A Big Big Data Platform

Similar documents
Building Bridges: A System for New HPC Communities

A Big Big Data Platform

The Data exacell DXC. J. Ray Scott DXC PI May 17, 2016

Our Workshop Environment

Our Workshop Environment

Our Workshop Environment

Our Workshop Environment

Our Workshop Environment

The Data Exacell (DXC): Data Infrastructure Building Blocks for Integrating Analytics with Data Management

Our Workshop Environment

Welcome to the XSEDE Big Data Workshop

Welcome to the XSEDE Big Data Workshop

Welcome to the XSEDE Big Data Workshop

A Converged HPC & Big Data Architecture in Production

HPC Cloud at SURFsara

A Breakthrough in Non-Volatile Memory Technology FUJITSU LIMITED

The GISandbox: A Science Gateway For Geospatial Computing. Davide Del Vento, Eric Shook, Andrea Zonca

Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21)

Comet Virtualization Code & Design Sprint

Un8l now, milestones in ar8ficial intelligence (AI) have focused on situa8ons where informa8on is complete, such as chess and Go.

A Converged HPC & Big Data System for Nontraditional and HPC Research

WVU RESEARCH COMPUTING INTRODUCTION. Introduction to WVU s Research Computing Services

ACCI Recommendations on Long Term Cyberinfrastructure Issues: Building Future Development

irods at TACC: Secure Infrastructure for Open Science Chris Jordan

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy

MOHA: Many-Task Computing Framework on Hadoop

CSD3 The Cambridge Service for Data Driven Discovery. A New National HPC Service for Data Intensive science

HIGH PERFORMANCE COMPUTING (PLATFORMS) SECURITY AND OPERATIONS

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science

Introducing SUSE Enterprise Storage 5

What s New at AWS? looking at just a few new things for Enterprise. Philipp Behre, Enterprise Solutions Architect, Amazon Web Services

Organizational Update: December 2015

LBRN - HPC systems : CCT, LSU

2013 AWS Worldwide Public Sector Summit Washington, D.C.

Predicting Service Outage Using Machine Learning Techniques. HPE Innovation Center

Abstract. The Challenges. ESG Lab Review InterSystems IRIS Data Platform: A Unified, Efficient Data Platform for Fast Business Insight

Journey Towards Science DMZ. Suhaimi Napis Technical Advisory Committee (Research Computing) MYREN-X Universiti Putra Malaysia

Making Supercomputing More Available and Accessible Windows HPC Server 2008 R2 Beta 2 Microsoft High Performance Computing April, 2010

IBM Power Systems Update. David Spurway IBM Power Systems Product Manager STG, UK and Ireland

The intelligence of hyper-converged infrastructure. Your Right Mix Solution

The Materials Data Facility

HPC learning using Cloud infrastructure

The iplant Data Commons

Building NVLink for Developers

EarthCube and Cyberinfrastructure for the Earth Sciences: Lessons and Perspective from OpenTopography

Data Movement & Tiering with DMF 7

Cisco and Cloudera Deliver WorldClass Solutions for Powering the Enterprise Data Hub alerts, etc. Organizations need the right technology and infrastr

Baremetal with Apache CloudStack

Emerging Technologies for HPC Storage

Big Data 2015: Sponsor and Participants Research Event ""

XSEDE s Campus Bridging Project Jim Ferguson National Institute for Computational Sciences

XSEDE Visualization Use Cases

Session 201-B: Accelerating Enterprise Applications with Flash Memory

Atos announces the Bull sequana X1000 the first exascale-class supercomputer. Jakub Venc

HPE SimpliVity. The new powerhouse in hyperconvergence. Boštjan Dolinar HPE. Maribor Lancom

Cisco Unified Computing System Delivering on Cisco's Unified Computing Vision

A Container On a Virtual Machine On an HPC? Presentation to HPC Advisory Council. Perth, July 31-Aug 01, 2017

THE NATIONAL DATA SERVICE(S) & NDS CONSORTIUM A Call to Action for Accelerating Discovery Through Data Services we can Build Ed Seidel

Storage Virtualization. Eric Yen Academia Sinica Grid Computing Centre (ASGC) Taiwan

Accelerate AI with Cisco Computing Solutions

The Stampede is Coming Welcome to Stampede Introductory Training. Dan Stanzione Texas Advanced Computing Center

DDN. DDN Updates. DataDirect Neworks Japan, Inc Nobu Hashizume. DDN Storage 2018 DDN Storage 1

Data publication and discovery with Globus

HCI: Hyper-Converged Infrastructure

SLATE. Services Layer at the Edge. First Meeting of the National Research Platform Montana State University August 7-8, 2017

XSEDE Campus Bridging Tools Rich Knepper Jim Ferguson

I/O at the Center for Information Services and High Performance Computing

The Canadian CyberSKA Project

Advanced Research Compu2ng Informa2on Technology Virginia Tech

Harp-DAAL for High Performance Big Data Computing

Jetstream: Adding Cloud-based Computing to the National Cyberinfrastructure

Cisco UCS B460 M4 Blade Server

Simplify Client Virtualization

Title DC Automation: It s a MARVEL!

PI SERVER 2012 Do. More. Faster. Now! Copyri g h t 2012 OSIso f t, LLC.

Jetstream Overview A national research and education cloud

New Approach to Unstructured Data

Dell EMC All-Flash solutions are powered by Intel Xeon processors. Learn more at DellEMC.com/All-Flash

HPE ProLiant ML350 Gen10 Server

Architectures for Scalable Media Object Search

Advancing Library Cyberinfrastructure for Big Data Sharing and Reuse. Zhiwu Xie

Minnesota Supercomputing Institute Regents of the University of Minnesota. All rights reserved.

IBM CORAL HPC System Solution

Interconnect Your Future

Cornell Red Cloud: Campus-based Hybrid Cloud. Steven Lee Cornell University Center for Advanced Computing

Modernizing Healthcare IT for the Data-driven Cognitive Era Storage and Software-Defined Infrastructure

Running VMware vsan Witness Appliance in VMware vcloudair First Published On: April 26, 2017 Last Updated On: April 26, 2017

DDN. DDN Updates. Data DirectNeworks Japan, Inc Shuichi Ihara. DDN Storage 2017 DDN Storage

Data Analytics and Storage System (DASS) Mixing POSIX and Hadoop Architectures. 13 November 2016

HPC Capabilities at Research Intensive Universities

UCS M-Series + Citrix XenApp Optimizing high density XenApp deployment at Scale

ADAC Federated Testbed Creating a Blueprint for Portable Ecosystems

Interconnect Your Future Enabling the Best Datacenter Return on Investment. TOP500 Supercomputers, November 2017

S8765 Performance Optimization for Deep- Learning on the Latest POWER Systems

University at Buffalo Center for Computational Research

The Future of Galaxy. Nate Coraor galaxyproject.org

Design patterns for data-driven research acceleration

HPE SimpliVity 380. Simplyfying Hybrid IT with HPE Wolfgang Privas Storage Category Manager

Transcription:

A Big Big Data Platform John Urbanic, Parallel Computing Scientist 2017 Pittsburgh Supercomputing Center

The Shift to Big Data New Emphases Pan-STARRS telescope http://pan-starrs.ifa.hawaii.edu/public/ Social networks and the Internet Video Wikipedia Commons Library of Congress stacks https://www.flickr.com/photos/danlem2001/6922113091/ Genome sequencers (Wikipedia Commons) NOAA climate modeling http://www.ornl.gov/info/ornlreview/v42_3_09/article02.shtml Collections Horniman museum: http://www.horniman.ac.uk/ get_involved/blog/bioblitz-insects-reviewed Legacy documents Wikipedia Commons Environmental sensors: Water temperature profiles from tagged hooded seals http://www.arctic.noaa.gov/report11/biodiv_whales_walrus.html 2

Challenges and Software are Co-Evolving Structured Data Machine Learning Unstructured Data Optimization (decision-making) Statistics Natural Language Processing Optimization (numerical) Calculations on Data Video Sound Image Analysis Scientific Visualization Graph Analytics Information Visualization 3

An Important Addition to the National Advanced Cyberinfrastructure Ecosystem Bridges will be a new resource on XSEDE and will interoperate with other XSEDE resources, Advanced Cyberinfrastructure (ACI) projects, campuses, and instruments nationwide. Examples: Reconstructing brain circuits from highresolution electron microscopy Carnegie Mellon University s Gates Center for Computer Science 4 High-throughput genome sequencers Social networks and the Internet Data Infrastructure Building Blocks (DIBBs) Data Exacell (DXC) Integrating Geospatial Capabilities into HUBzero Building a Scalable Infrastructure for Data- Driven Discovery & Innovation in Education Other DIBBs projects Other ACI projects Temple University s new Science, Education, and Research Center

Motivating Use Cases (examples) Data-intensive applications workflows Gateways the power of HPC without the programming Shared data collections related analysis tools Cross-domain analytics } Graph analytics, machine learning, genome sequence assembly, and other large-memory applications Scaling research questions beyond the laptop Scaling research from individuals to teams and collaborations Very large in-memory databases Optimization Distributed parameter sweeps service-oriented architectures Data assimilation from large instruments and Internet data Leveraging an extensive collection of interoperating software Research areas that haven t used HPC Nontraditional HPC approaches to fields such as the physical sciences Coupling applications in novel ways Leveraging large memory and high island bandwidth 5

Objectives and Approach Bring HPC to nontraditional users and research communities. Allow high-performance computing to be applied effectively to big data. Bridge to campuses to streamline access and provide cloud-like burst capability. Leveraging PSC s expertise with shared memory, Bridges will feature 3 tiers of large, coherent shared-memory nodes: 12TB, 3TB, and 128GB. EMBO Mol Med (2013) DOI: 10.1002/emmm.201202388: Proliferation of cancer-causing mutations throughout life Alex Hauptmann et. al.: Efficient large-scale content-based multimedia event detection Bridges implements a uniquely flexible environment featuring interactivity, gateways, databases, distributed (web) services, high-productivity programming languages and frameworks, and virtualization, and campus bridging. 6

Interactivity Interactivity is the feature most frequently requested by nontraditional HPC communities. Interactivity provides immediate feedback for doing exploratory data analytics and testing hypotheses. Bridges offers interactivity through a combination of virtualization for lighter-weight applications and dedicated nodes for more demanding ones. 7

Gateways and Tools for Building Them Gateways provide easy-to-use access to Bridges HPC and data resources, allowing users to launch jobs, orchestrate complex workflows, and manage data from their browsers. - Extensive leveraging of databases and polystore systems - Great attention to HCI is needed to get these right Interactive pipeline creation in GenePattern (Broad Institute) Col*Fusion portal for the systematic accumulation, integration, and utilization of historical data, from http://colfusion.exp.sis.pitt.edu/colfusion/ Download sites for MEGA-6 (Molecular Evolutionary Genetic Analysis), from www.megasoftware.net 8

Virtualization and Containers Virtual Machines (VMs) enable flexibility, security, customization, reproducibility, ease of use, and interoperability with other services. User demand is for custom database and web server installations to develop data-intensive, distributed applications and containers for custom software stacks and portability. Bridges leverages OpenStack to provision resources, between interactive, batch, Hadoop, and VM uses. 9

High-Productivity Programming Supporting languages that communities already use is vital for them to apply HPC to their research questions. 10

Spark, Hadoop Related Approaches Bridges large memory is great for Spark! Bridges enables workflows that integrate Spark/Hadoop, HPC, and/or shared-memory components. 11

Campus Bridging Federated identity management Burst offload http://www.temple.edu/medicine/research/research_tusm/ Through a pilot project with Temple University, the Bridges project will explore new ways to transition data and computing seamlessly between campus and XSEDE resources. Federated identity management will allow users to use their local credentials for single sign-on to remote resources, facilitating data transfers between Bridges and Temple s local storage systems. Burst offload will enable cloud-like offloading of jobs from Temple to Bridges and vice versa during periods of unusually heavy load. 12

20 Storage Building Blocks, implementing the parallel Pylon filesystem (~10PB) using PSC s SLASH2 filesystem 4 MDS nodes 2 front-end nodes 2 boot nodes 8 management nodes 6 core Intel OPA edge switches: fully interconnected, 2 links per switch Intel OPA cables 4 HPE Integrity Superdome X (12TB) compute nodes 42 HPE ProLiant DL580 (3TB) compute nodes 12 HPE ProLiant DL380 database nodes 6 HPE ProLiant DL360 web server nodes 20 leaf Intel OPA edge switches 32 RSM nodes with NVIDIA next-generation GPUs 800 HPE Apollo 2000 (128GB) compute nodes 13 2015 Pittsburgh Supercomputing Center 16 RSM nodes with NVIDIA K80 GPUs Purpose-built Intel Omni-Path topology for data-intensive HPC http://staff.psc.edu/nystrom/bvt http://psc.edu/bridges

Node Types 14

Database and Web Server Nodes Dedicated database nodes will power persistent relational and NoSQL databases Support data management and data-driven workflows SSDs for high IOPs; RAIDed HDDs for high capacity (examples) Dedicated web server nodes Enable distributed, service-oriented architectures High-bandwidth connections to XSEDE and the Internet 15

GPU Nodes Bridges GPUs are accelerating both deep learning and simulation codes Phase 1: 16 nodes, each with: 2 Intel Xeon E5-2695 v3 (14c, 2.3/3.3 GHz) 128GB DDR4-2133 RAM Phase 2: +32 nodes, each with: 2 Intel Xeon E5-2683 v4 (16c, 2.1/3.0 GHz) 128GB DDR4-2400 RAM 16

Data Management Pylon: A large, central, high-performance filesystem Visible to all nodes Large datasets, community repositories (10 PB usable) Distributed (node-local) storage Enhance application portability Improve overall system performance Improve performance consistency to the shared filesystem Acceleration for Hadoop-based applications 17

Intel Omni-Path Architecture (OPA) Bridges is the first production deployment of Omni-Path Omni-Path connects all nodes and the shared filesystem, providing Bridges and its users with: 100 Gbps line speed per port; 25 GB/s bidirectional bandwidth per port Observed <0.93μs latency, 12.36 GB/s/dir 160M MPI messages per second 48-port edge switch reduces interconnect complexity and cost HPC performance, reliability, and QoS OFA-compliant applications supported without modification Early access to this new, important, forward-looking technology Bridges deploys OPA in a two-tier island topology developed by PSC for cost-effective, data-intensive HPC 18

Example: Causal Discovery Portal Center for Causal Discovery, an NIH Big Data to Knowledge Center of Excellence Browser-based UI Prepare and upload data Run causal Internet discovery algorithms Visualize results Web node VM Apache Tomcat Messaging Database node VM Authentication Data Provenance Other DBs Execute causal discovery algorithms Omni-Path Pylon filesystem TCGA fmri Analytics: FGS and other algorithms, building on TETRAD LSM Node (3TB) Memoryresident datasets ESM Node (12TB) 19

Gaining Access to Bridges Bridges is allocated through XSEDE: https://www.xsede.org/allocations Starter Allocation Can request anytime Up to 50,000 core-hours on RSM and GPU (128GB) nodes and/or 10,000 TB-hours on LSM (3TB) and ESM (12TB) nodes Can request XSEDE ECSS (Extended Collaborative Support Service Research Allocation (XRAC) Appropriate for larger requests; can request ECSS Can be up to millions to tens of millions of SUs Quarterly submission windows: March 15 April 15, June 15 July 15, etc. Coursework Allocations To support use of Bridges for relevant courses Community Allocations Primarily to support gateways XSEDE calls these Service Units (SUs) Up to 10% of Bridges SUs are available on a discretionary basis to industrial affiliates, Pennsylvania-based researchers, and others to foster discovery and innovation and broaden participation in data-intensive computing. 20

For Additional Information Project website: www.psc.edu/bridges Questions: bridges@psc.edu Bridges PI: Nick Nystrom nystrom@psc.edu Co-PIs: Michael J. Levine Ralph Roskies J Ray Scott Project Manager: Robin Scibek 21