Building Bridges: A System for New HPC Communities

Similar documents
A Big Big Data Platform

A Big Big Data Platform

The Data exacell DXC. J. Ray Scott DXC PI May 17, 2016

Our Workshop Environment

The Data Exacell (DXC): Data Infrastructure Building Blocks for Integrating Analytics with Data Management

Our Workshop Environment

Our Workshop Environment

Our Workshop Environment

Our Workshop Environment

Welcome to the XSEDE Big Data Workshop

Our Workshop Environment

Welcome to the XSEDE Big Data Workshop

Welcome to the XSEDE Big Data Workshop

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

ACCI Recommendations on Long Term Cyberinfrastructure Issues: Building Future Development

HPC Cloud at SURFsara

A Breakthrough in Non-Volatile Memory Technology FUJITSU LIMITED

CSD3 The Cambridge Service for Data Driven Discovery. A New National HPC Service for Data Intensive science

WVU RESEARCH COMPUTING INTRODUCTION. Introduction to WVU s Research Computing Services

Co-existence: Can Big Data and Big Computation Co-exist on the Same Systems?

MOHA: Many-Task Computing Framework on Hadoop

Comet Virtualization Code & Design Sprint

Cisco UCS B460 M4 Blade Server

HIGH PERFORMANCE COMPUTING (PLATFORMS) SECURITY AND OPERATIONS

2013 AWS Worldwide Public Sector Summit Washington, D.C.

Advanced Research Compu2ng Informa2on Technology Virginia Tech

A Converged HPC & Big Data Architecture in Production

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

DDN. DDN Updates. DataDirect Neworks Japan, Inc Nobu Hashizume. DDN Storage 2018 DDN Storage 1

LBRN - HPC systems : CCT, LSU

A Converged HPC & Big Data System for Nontraditional and HPC Research

XSEDE s Campus Bridging Project Jim Ferguson National Institute for Computational Sciences

The GISandbox: A Science Gateway For Geospatial Computing. Davide Del Vento, Eric Shook, Andrea Zonca

Baremetal with Apache CloudStack

Organizational Update: December 2015

The Materials Data Facility

Atos announces the Bull sequana X1000 the first exascale-class supercomputer. Jakub Venc

Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21)

Journey Towards Science DMZ. Suhaimi Napis Technical Advisory Committee (Research Computing) MYREN-X Universiti Putra Malaysia

Predicting Service Outage Using Machine Learning Techniques. HPE Innovation Center

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy

Gen-Z Memory-Driven Computing

irods at TACC: Secure Infrastructure for Open Science Chris Jordan

Exam C Foundations of IBM Cloud Reference Architecture V5

Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science

Storage Virtualization. Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan

DDN. DDN Updates. Data DirectNeworks Japan, Inc Shuichi Ihara. DDN Storage 2017 DDN Storage

Harp-DAAL for High Performance Big Data Computing

THE API DEVELOPER EXPERIENCE ENABLING RAPID INTEGRATION

X-ray imaging software tools for HPC clusters and the Cloud

Un8l now, milestones in ar8ficial intelligence (AI) have focused on situa8ons where informa8on is complete, such as chess and Go.

Abstract. The Challenges. ESG Lab Review InterSystems IRIS Data Platform: A Unified, Efficient Data Platform for Fast Business Insight

The Stampede is Coming Welcome to Stampede Introductory Training. Dan Stanzione Texas Advanced Computing Center

TESLA P100 PERFORMANCE GUIDE. HPC and Deep Learning Applications

SuperMike-II Launch Workshop. System Overview and Allocations

THE NATIONAL DATA SERVICE(S) & NDS CONSORTIUM A Call to Action for Accelerating Discovery Through Data Services we can Build Ed Seidel

The Cambridge Bio-Medical-Cloud An OpenStack platform for medical analytics and biomedical research

Design patterns for data-driven research acceleration

AN INTRODUCTION TO CLUSTER COMPUTING

The iplant Data Commons

Software as a Service Gateways

UCS M-Series + Citrix XenApp Optimizing high density XenApp deployment at Scale

Smart Trading with Cray Systems: Making Smarter Models + Better Decisions in Algorithmic Trading

2013 Cisco and/or its affiliates. All rights reserved. 1

XSEDE Visualization Use Cases

IBM CORAL HPC System Solution

RIDESHARING YOUR CLOUD DATA CENTER Realize Better Resource Utilization with NVMe-oF

Oracle Big Data Connectors

TECHNICAL OVERVIEW ACCELERATED COMPUTING AND THE DEMOCRATIZATION OF SUPERCOMPUTING

Cloud Computing For Researchers

The Future of Galaxy. Nate Coraor galaxyproject.org

Building NVLink for Developers

Flash Storage Complementing a Data Lake for Real-Time Insight

HPC Enabling R&D at Philip Morris International

Jetstream: Adding Cloud-based Computing to the National Cyberinfrastructure

RIDESHARING YOUR CLOUD DATA CENTER Realize Better Resource Utilization with NVMe-oF

DDN About Us Solving Large Enterprise and Web Scale Challenges

Modernizing Healthcare IT for the Data-driven Cognitive Era Storage and Software-Defined Infrastructure

IBM Power Systems Update. David Spurway IBM Power Systems Product Manager STG, UK and Ireland

Regional & National HPC resources available to UCSB

University at Buffalo Center for Computational Research

Data Movement & Tiering with DMF 7

AWS & Intel: A Partnership Dedicated to fueling your Innovations. Thomas Kellerer BDM CSP, Intel Central Europe

EarthCube and Cyberinfrastructure for the Earth Sciences: Lessons and Perspective from OpenTopography

Accelerate OpenStack* Together. * OpenStack is a registered trademark of the OpenStack Foundation

NUIT Tech Talk Topics in Research Computing: XSEDE and Northwestern University Campus Champions

THE DEFINITIVE GUIDE FOR AWS CLOUD EC2 FAMILIES

The BioHPC Nucleus Cluster & Future Developments

TACC s Stampede Project: Intel MIC for Simulation and Data-Intensive Computing

VxRail: Level Up with New Capabilities and Powers GLOBAL SPONSORS

IBM Power Systems HPC Cluster

What s New at AWS? looking at just a few new things for Enterprise. Philipp Behre, Enterprise Solutions Architect, Amazon Web Services

Scaling Internet TV Content Delivery ALEX GUTARIN DIRECTOR OF ENGINEERING, NETFLIX

Transform to Your Cloud

The intelligence of hyper-converged infrastructure. Your Right Mix Solution

Modern Data Warehouse The New Approach to Azure BI

Big Data 2015: Sponsor and Participants Research Event ""

Data publication and discovery with Globus

in Action Fujitsu High Performance Computing Ecosystem Human Centric Innovation Innovation Flexibility Simplicity

Xcellis Technical Overview: A deep dive into the latest hardware designed for StorNext 5

Transcription:

Building Bridges: A System for New HPC Communities HPC User Forum 59 LRZ, Garching October 16, 2015 Presenter: Jim Kasdorf Director, Special Projects Pittsburgh Supercomputing Center kasdorf@psc.edu 2015 Pittsburgh Supercomputing Center

A Uniquely Flexible HPC Resource for New Communities and Data Analytics Nick Nystrom Director, Strategic Applications & Bridges PI nystrom@psc.edu October 7, 2015 2015 Pittsburgh Supercomputing Center

The Shift to Big Data New Emphases Pan-STARRS telescope http://pan-starrs.ifa.hawaii.edu/public/ Social networks and the Internet Video Wikipedia Commons Library of Congress stacks https://www.flickr.com/photos/danlem2001/6922113091/ Genome sequencers (Wikipedia Commons) NOAA climate modeling http://www.ornl.gov/info/ornlreview/v42_3_09/article02.shtml Collections Horniman museum: http://www.horniman.ac.uk/ get_involved/blog/bioblitz-insects-reviewed Legacy documents Wikipedia Commons Environmental sensors: Water temperature profiles from tagged hooded seals http://www.arctic.noaa.gov/report11/biodiv_whales_walrus.html 2

Algorithms and Applications Algorithms and Applications Have Also Changed Structured Data Machine Learning Unstructured Data Optimization (decision-making) Statistics Natural Language Processing Optimization (numerical) Video Image Analysis Calculations on Data Sound Scientific Visualization Graph Analytics Information Visualization 3

The $9.65M Bridges acquisition is made possible by National Science Foundation Bridges: From Communities and Data to Workflows and Insight HPE is delivering Bridges Disclaimer: The following presentation conveys the current plan for Bridges. Details are subject to change. 4

High-Level Use Cases Data-intensive applications & workflows Gateways the power of HPC without the programming Shared data collections & related analysis tools Crossdomain analytics Graph analytics, machine learning, genome sequence assembly, and other large-memory applications Scaling research questions beyond the laptop Scaling research from individuals to teams and collaborations Very large in-memory databases Optimization & parameter sweeps Distributed & service-oriented architectures Data assimilation from large instruments and Internet data Leveraging an extensive collection of interoperating software 8

Potential Applications (Examples) Finding causal relationships in cancer genomics, lung disease, and brain dysfunction Analysis of financial markets and policies Improving the effectiveness of organ donation networks Assembling large genomes and metagenomes Recognizing events and enabling search for videos Understanding how the brain is connected from EM data Addressing societal issues from social media data Analyzing large corpora in the digital humanities Cross-observational analyses in astronomy & other sciences Data integration and fusion for history and related fields 9

Objectives and Approach Bring HPC to nontraditional users and research communities. Allow high-performance computing to be applied effectively to big data. Bridge to campuses to ease access and provide burst capability. Leveraging PSC s expertise with shared memory, Bridges will feature 3 tiers of large, coherent shared-memory nodes. EMBO Mol Med (2013) DOI: 10.1002/emmm.201202388: Proliferation of cancer-causing mutations throughout life Alex Hauptmann et. al.: Efficient large-scale content-based multimedia event detection Bridges will leverage its large memory for interactivity and seamlessly support applications through virtualization, gateways, familiar and productive programming environments and data-driven workflows. 10

User-Friendly HPC & Data Analytics Interactivity is the feature most frequently requested by nontraditional HPC communities and for doing data analytics and testing hypotheses. Gateways and tools for gateway building will provide easy-to-use access to Bridges HPC and data resources. Database and web server nodes will provide persistent databases to enable data management, workflows, and distributed applications. Popular programming languages & software environments will let users scale applications and workflows. Virtualization will allow users to bring their particular environments and provide interoperability with clouds. 11

Interactivity Interactivity is the feature most frequently requested by nontraditional HPC communities. Interactivity provides immediate feedback for doing exploratory data analytics and testing hypotheses. Bridges will offer interactivity through a combination of virtualization for lighter-weight applications and dedicated nodes for more demanding ones. 12

Gateways and Tools for Building Them Gateways will provide easy-to-use access to Bridges HPC and data resources, allowing users to launch jobs, orchestrate complex workflows and manage data from their web browsers. Interactive pipeline creation in GenePattern (Broad Institute) Col*Fusion portal for the systematic accumulation, integration, and utilization of historical data, from http://colfusion.exp.sis.pitt.edu/colfusion/ Download sites for MEGA-6 (Molecular Evolutionary Genetic Analysis), from www.megasoftware.net 13

Virtualization and Containers Virtual Machines (VMs) will enable flexibility, customization, security, reproducibility, ease of use, and interoperability with other services. Early user demand on PSC s Data Exacell (a research pilot project) has centered on VMs for custom database and web server installations to develop data-intensive, distributed applications and containers for reproducibility. We plan to leverage OpenStack to provision resources, between interactive, batch, Hadoop, and VM uses. 14

High-Productivity Programming Bridges will feature high-productivity programming languages and tools. 15

Hadoop Ecosystem Bridges will provide acceleration for Hadoop applications Large memory will be great for Spark. 16

Campus Bridging Federated identity management Burst offload http://www.temple.edu/medicine/research/research_tusm/ Through a pilot project with Temple University, the Bridges project will explore new ways to transition data and computing seamlessly between campus and XSEDE resources. Federated identity management will allow users to use their local credentials for single sign-on to remote resources, facilitating data transfers between Bridges and Temple s local storage systems. Burst offload will enable cloud-like offloading of jobs from Temple to Bridges and vice versa during periods of unusually heavy load. Custom PSC topology for data-intensive HPC 20 Storage Building Blocks, implementing

the parallel Pylon filesystem (~10PB) using PSC s SLASH2 filesystem 4 MDS nodes 2 front-end nodes 2 boot nodes 8 management nodes 6 core Intel OPA edge switches: fully interconnected, 2 links per switch Intel OPA cables 4 ESM (12 TB) compute nodes 2 gateways per ESM 42 LSM (3 TB) compute nodes 12 database nodes 6 web server nodes 20 leaf Intel OPA edge switches 32 RSM nodes with NVIDIA next-generation GPUs 800 RSM (128 GB) compute nodes, 48 with GPUs 16 RSM nodes with NVIDIA K80 GPUs 19

High-Performance, Data-Intensive Computing Three tiers of large, coherent shared memory nodes Memory per node Number of nodes Example applications 12 TB Several Genomics, machine learning, graph analytics, other extreme-memory applications 3 TB Tens Virtualization and interactivity including large-scale visualization and analytics; mid-spectrum memory-intensive jobs 128 GB Hundreds Execution of most components of workflows, interactivity, Hadoop, and capacity computing The latest Intel Xeon CPUs NVIDIA Tesla dual-gpu accelerators 20

Database and Web Server Nodes Dedicated database nodes will power persistent relational and NoSQL databases Support data management and data-driven workflows SSDs for high IOPs; RAIDed HDDs for high capacity (examples) Dedicated web server nodes Enable distributed, service-oriented architectures High-bandwidth connections to XSEDE and the Internet 21

Data Management Pylon: A large, central, high-performance filesystem Visible to all nodes Large datasets, community repositories (~10 PB usable) Distributed (node-local) storage Enhance application portability Improve overall system performance Improve performance consistency to the shared filesystem Acceleration for Hadoop-based applications 22

Intel Omni-Path Architecture Omni-Path will connect all nodes and the shared filesystem, providing Bridges and its users with: 100 Gbps line speed per port; 25 GB/s bidirectional bandwidth per port 160M MPI messages per second 48-port edge switch reduces interconnect complexity and cost HPC performance, reliability, and QoS OFA-compliant applications supported without modification Early access to this new, important, forward-looking technology The Intel Omni-Path Architecture will deployed in Bridges using a novel topology developed by PSC to enable dataintensive HPC 23

Example: Causal Discovery Browser-based UI Prepare and upload data Run causal discovery algorithms Visualize results Internet Web node VM Apache Tomcat Messaging Database node VM Authentication Data Provenance Other DBs Execute causal discovery algorithms Omni-Path Pylon filesystem TCGA fmri Analytics: TETRAD (FastGES, ) and related implementations LSM Node (3TB) Memoryresident datasets ESM Node (12TB) 24

Getting Started on Bridges Starter Allocation https://www.xsede.org/allocations Can request anytime... including now! 1-year effective duration, to begin when Bridges comes online; can begin running on Greenfield Can request XSEDE ECSS (Extended Collaborative Support Service) Research Allocation (XRAC) https://www.xsede.org/allocations Appropriate for larger requests Quarterly submission windows; Next: Sept. 15 Oct. 15, 2015 Can request ECSS Early User Period Users with starter or research proposals may be eligible for Bridges Early User Period (late fall 15) 25

Bridges Target Schedule Acquisition Construction planned to begin October, 2015 Early User Period starting in late 2015 XRAC Allocated Use Planned to begin in January, 2016 Related resources Greenfield to transition from Blacklight and to provide data for developing XRAC proposals for Bridges Data Exacell: a research pilot project to explore the coupling of data analytics with novel storage 26

For Additional Information Project website: www.psc.edu/bridges Bridges PI: Nick Nystrom nystrom@psc.edu Co-PIs: Michael J. Levine Ralph Roskies J Ray Scott Project Manager: Janet Brown 27