The Lattice BOINC Project Public Computing for the Tree of Life

Similar documents
GPU computing and the tree of life

ML phylogenetic inference and GARLI. Derrick Zwickl. University of Arizona (and University of Kansas) Workshop on Molecular Evolution 2015

Condor and BOINC. Distributed and Volunteer Computing. Presented by Adam Bazinet

Managing CAE Simulation Workloads in Cluster Environments

Grid Mashups. Gluing grids together with Condor and BOINC

Grid Scheduling Architectures with Globus

Lecture 1: Introduction and Computational Thinking

Supercomputing resources at the IAC

The BOINC Community. PC volunteers (240,000) Projects. UC Berkeley developers (2.5) Other volunteers: testing translation support. Computer scientists

NVIDIA s Compute Unified Device Architecture (CUDA)

NVIDIA s Compute Unified Device Architecture (CUDA)

SEASHORE / SARUMAN. Short Read Matching using GPU Programming. Tobias Jakobi

Using Graphics Chips for General Purpose Computation

Introduction to Parallel Programming

High Performance Computing on GPUs using NVIDIA CUDA

Accelerating Molecular Modeling Applications with Graphics Processors

Approaches to Parallel Computing

Computing on GPUs. Prof. Dr. Uli Göhner. DYNAmore GmbH. Stuttgart, Germany

Parallel & Cluster Computing. cs 6260 professor: elise de doncker by: lina hussein

Real Parallel Computers

The rcuda middleware and applications

University at Buffalo Center for Computational Research

Phylogenetics on CUDA (Parallel) Architectures Bradly Alicea

CUDA Web API. Remote Execution of CUDA Kernels using Web Services. A Thesis. Presented To. The Faculty of California Polytechnic State University

THEBES: THE GRID MIDDLEWARE PROJECT Project Overview, Status Report and Roadmap

high performance medical reconstruction using stream programming paradigms

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

High-Performance and Parallel Computing

Enabling Phylogenetic Research via the CIPRES Science Gateway!

Future Trends in Hardware and Software for use in Simulation

The University of Oxford campus grid, expansion and integrating new partners. Dr. David Wallom Technical Manager

The Use of Cloud Computing Resources in an HPC Environment

Implementation of Deep Convolutional Neural Net on a Digital Signal Processor

GPU Clouds IGT Cloud Computing Summit Mordechai Butrashvily, CEO 2009 (c) All rights reserved

GROMACS Performance Benchmark and Profiling. August 2011

SuperMike-II Launch Workshop. System Overview and Allocations

Finite Difference Time Domain (FDTD) Simulations Using Graphics Processors

Lecture 1: January 22

Regional & National HPC resources available to UCSB

Ioan Raicu. Everyone else. More information at: Background? What do you want to get out of this course?

ACCELERATED COMPLEX EVENT PROCESSING WITH GRAPHICS PROCESSING UNITS

HPC Enabling R&D at Philip Morris International

Introduction to Cluster Computing

Advanced School in High Performance and GRID Computing November Introduction to Grid computing.

SWsoft ADVANCED VIRTUALIZATION AND WORKLOAD MANAGEMENT ON ITANIUM 2-BASED SERVERS

OpenCL Base Course Ing. Marco Stefano Scroppo, PhD Student at University of Catania

Experiences with GPGPUs at HLRS

Forget about the Clouds, Shoot for the MOON

G P G P U : H I G H - P E R F O R M A N C E C O M P U T I N G

Introduction to Grid Computing

GPU Programming Using NVIDIA CUDA

Introduction to High-Performance Computing

Chapter 3. Design of Grid Scheduler. 3.1 Introduction

Lecture 1: January 23

Gene Sequence Alignment on a Public Computing Platform

Perform scalable data exchange using InfoSphere DataStage DB2 Connector

Cycle Sharing Systems

NFS, GPFS, PVFS, Lustre Batch-scheduled systems: Clusters, Grids, and Supercomputers Programming paradigm: HPC, MTC, and HTC

A320 Supplemental Multi-Core Materials

Day 9: Introduction to CHTC

BlueGene/L (No. 4 in the Latest Top500 List)

Fahad Zafar, Dibyajyoti Ghosh, Lawrence Sebald, Shujia Zhou. University of Maryland Baltimore County

HPC with GPU and its applications from Inspur. Haibo Xie, Ph.D

INSTITUTO SUPERIOR TÉCNICO. Architectures for Embedded Computing

Outline. Definition of a Distributed System Goals of a Distributed System Types of Distributed Systems

Corral: A Glide-in Based Service for Resource Provisioning

PBS PROFESSIONAL VS. MICROSOFT HPC PACK

GPU for HPC. October 2010

NCAR Globally Accessible Data Environment (GLADE) Updated: 15 Feb 2017

Overview and Introduction to Scientific Visualization. Texas Advanced Computing Center The University of Texas at Austin

How to Write Fast Code , spring th Lecture, Mar. 31 st

EGEE and Interoperation

Real Parallel Computers

Clustering. Research and Teaching Unit

Grand Central Dispatch

High-throughput Sequence Alignment using Graphics Processing Units

GPU 101. Mike Bailey. Oregon State University. Oregon State University. Computer Graphics gpu101.pptx. mjb April 23, 2017

GPU 101. Mike Bailey. Oregon State University

General Purpose GPU Computing in Partial Wave Analysis

WHY PARALLEL PROCESSING? (CE-401)

Maximize automotive simulation productivity with ANSYS HPC and NVIDIA GPUs

A MULTI-GPU COMPUTE SOLUTION FOR OPTIMIZED GENOMIC SELECTION ANALYSIS. A Thesis. presented to. the Faculty of California Polytechnic State University

Large scale Imaging on Current Many- Core Platforms

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

ECE 574 Cluster Computing Lecture 16

The BioHPC Nucleus Cluster & Future Developments

GPUs and GPGPUs. Greg Blanton John T. Lubia

TOOLS FOR IMPROVING CROSS-PLATFORM SOFTWARE DEVELOPMENT

Introduction to GPU hardware and to CUDA

Interpreting a genetic programming population on an nvidia Tesla

Complexity and Advanced Algorithms. Introduction to Parallel Algorithms

Introduction to computing, architecture and the UNIX OS. HORT Lecture 1 Instructor: Kranthi Varala

Shaking-and-Baking on a Grid

Managing data flows. Martyn Winn Scientific Computing Dept. STFC Daresbury Laboratory Cheshire. 8th May 2014

Chapter 2 Parallel Hardware

The Grid: Feng Shui for the Terminally Rectilinear

A CAM(Content Addressable Memory)-based architecture for molecular sequence matching

WVU RESEARCH COMPUTING INTRODUCTION. Introduction to WVU s Research Computing Services

AN INTRODUCTION TO CLUSTER COMPUTING

Summer 2009 REU: Introduction to Some Advanced Topics in Computational Mathematics

Transcription:

The Lattice BOINC Project Public Computing for the Tree of Life Presented by Adam Bazinet Center for Bioinformatics and Computational Biology Institute for Advanced Computer Studies University of Maryland

Parallel computing universe Multi-core GPU SMP Cluster MPP Grid computing

Parallel computing universe Shared memory Multi-core GPU SMP

Parallel computing universe Multi-core GPU SMP Cluster MPP Grid computing

Parallel computing universe Distributed memory Cluster MPP Grid computing

Parallel computing universe Distributed computing Cluster MPP Grid computing

Parallel computing universe Multi-core GPU SMP Cluster MPP Grid computing

Parallel computing universe GPU SMP Cluster MPP Multi-core Grid computing

Grid computing BOINC pool Condor pool Compute cluster

The Lattice Project: resource characterization HTC resources HPC resources Condor pool BOINC pool Compute cluster HTC = high throughput computing HPC = high performance computing

The Lattice Project: resource characterization Shared Dedicated Condor pool BOINC pool Compute cluster shared = resources usually have interactive users dedicated = resources dedicated to processing

The Lattice Project: resource characterization Institutional Volunteer Condor pool Compute cluster BOINC pool institutional = resources belong to an institution dedicated = resources volunteered by the public

The Lattice Project: resource characterization Trusted Untrusted Condor pool Compute cluster BOINC pool trusted = results need not be verified untrusted = results need verification

The Lattice Project: resource characterization Reliable Unreliable Condor pool Compute cluster BOINC pool reliable = resources are stable and accounted for unreliable = resources are dynamic and unpredictable

Models of Grid computing SERVICE MODEL a heavyweight, feature-rich model focused on providing access to institutional resources and robust job submission capabilities and security features Well known Service Grids include TeraGrid, Open Science Grid, and EGEE DESKTOP MODEL scavenges cycles from idle desktop computers, which are volunteered by the general public The combined power of hundreds of thousands of desktop computers represents a substantial, readily available resource The most widely used software for tapping this resource is BOINC

The first Grid system to effectively combine a Service Grid (using Globus software) and a Desktop Grid (using BOINC software) Aimed at sharing computational resources between academic institutions, particularly those in the University System of Maryland Focused on enabling large-scale computation, especially for problems in the life sciences Development began in 2003 since then, many different researchers have used the system, racking up over 20,000 CPU years of computation (measured in wall clock time)

The Lattice Project: architecture Condor Central Manager Condor Compute Nodes

The Lattice Project: resources Quick facts about resources: We support three major platforms: Linux (both PowerPC and Intel-based), Windows, and Mac (both PowerPC and Intel-based) Four different institutions are currently tied in to the Grid: UMCP, Bowie State University, Coppin State University, and the Smithsonian NMNH We currently have four Condor pools, four dedicated clusters, and a BOINC project with a steady number of participants We currently have a total of 4000-5000 CPUs

The Lattice Project: resources http://lattice.umiacs.umd.edu/resources/

The Lattice BOINC Project http://boinc.umiacs.umd.edu/

BOINC as part of a Grid computing system The Lattice BOINC Project is part of a Grid computing system that also incorporates traditional computing clusters and Condor pools How it works: a job with a generic description is submitted by a Grid user, which is transformed into either a BOINC workunit, a PBS submit file, or a Condor submit file by the Globus GRAM service As a consequence of being hooked up to a general-purpose Grid computing system, our project runs many different types of applications As a result, we have had to work harder to manage volunteer expectations since we do not have an endless supply of work and the applications (and workunits within an application) are highly variable

The Lattice BOINC Project: research Phylogenetic analysis GARLI Protein sequence comparison HMMPfam Conservation network design MARXAN

Grid, Public, and GPU Computing for the Tree of Life A multi-year NSF award to build an advanced computational system for phylogenetic analysis Leverages the Grid system (and hence the Lattice BOINC Project) Provides for improving the performance of popular phylogenetic analysis programs (GARLI) using GPGPU frameworks such as OpenCL and CUDA The BOINC pool is our greatest potential source of modern GPUs Provides for the construction of a web portal interface to facilitate easy and efficient job submission, monitoring, and post-processing

GARLI Genetic Algorithm for Rapid Likelihood Inference Makes maximum likelihood phylogenetic searches feasible for large datasets Avoids calculating the exact maximized likelihood for each topology Gradually and independently optimizes tree topology, branch lengths, and model parameters

GPUs: background GPUs are inexpensive, dedicated numerical processors originally designed for rendering 3D computer graphics GPUs contain 100s of processing cores on a single chip Each core carries out the same operations in parallel on different input data single program, multiple data (SPMD) paradigm Extremely high arithmetic intensity if one can transfer the data onto and results off of the processors efficiently

GPUs: background general-purpose computing on GPUs NVIDIA GPU Intel CPU 1000 peak GFLOPS 750 500 250 2003 2004 2005 2006 2007 2008 NVIDIA Corporation. CUDA Programming Guide 2.1, 2008

The Lattice BOINC Project: hosts by country host count GPU-enabled period of January 1st to August 23rd, 2010; total of 15081 hosts (5750 GPU-enabled); created on Many Eyes IBM

The Lattice BOINC Project: GPU models period of January 1st to August 23rd, 2010; total of 5750 GPUs; created on Many Eyes IBM

GARLI-GPU: speedups nucleotide 2178 bp 11 taxa CPU only hosts CPU + GPU hosts amino acid 2178 bp 11 taxa codon 2178 bp 11 taxa nucleotide 18792 bp 15 taxa amino acid 18792 bp 15 taxa codon 18792 bp 15 taxa 0 2 4 6 8 speedup 1568 hosts (271 GPU-enabled); test from 2009

GARLI on BOINC: challenges and future work Checkpointing - GARLI only checkpoints in certain phases Progress bar updating - due to the stochastic GA and flexible stopping conditions, it can be difficult to ascertain progress Runtime estimates - providing accurate runtime estimates a priori is important for a variety of reasons; has been difficult to do so because of the high variance in types of workunits solution: machine learning with random forest to estimate runtimes in advance Long jobs, short pieces - if we could break up workunits into shorter, even-length pieces, we could give our volunteers a more consistent experience, and would probably increase throughput

References Bazinet, A. L. 2009. The Lattice Project: A Multi-model Grid Computing System. Master's thesis, University of Maryland. Bazinet, A. L., and M. P. Cummings. 2008. The Lattice Project: a Grid research and production environment combining multiple Grid computing models. Pages 2-13. In Weber, M. H. W. (Ed.) Distributed & Grid Computing - Science Made Transparent for Everyone. Principles, Applications and Supporting Communities. Rechenkraft.net, Marburg. Myers, D. S., A. L. Bazinet and M. P. Cummings. 2008. Expanding the reach of Grid computing: combining Globus- and BOINC-based systems. Pages 71-85. In Talbi, E.-G. and A. Zomaya (Eds.) Grids for Bioinformatics and Computational Biology, Wiley Book Series on Parallel and Distributed Computing. John Wiley & Sons, New York. Bazinet, A. L., D. S. Myers, J. Fuetsch and M. P. Cummings. 2007. Grid Services Base Library: a high-level, procedural application program interface for writing Globus-based Grid services. Future Generation Computer Systems 22:517-522. Myers, D. S., and M. P. Cummings. 2003. Necessity is the mother of invention: a simple Grid computing system using commodity tools. Journal of Parallel and Distributed Computing. 63:578-589. Zwickl, D. J. 2006. Genetic algorithm approaches for the phylogenetic analysis of large biological sequence datasets under the maximum likelihood criterion. Ph.D. dissertation, The University of Texas at Austin.

More information http://lattice.umiacs.umd.edu/ http://boinc.umiacs.umd.edu/ Contacts: - Adam Bazinet (pknut777@umiacs.umd.edu) - Michael Cummings (mike@umiacs.umd.edu)