Stream Processing for Remote Collaborative Data Analysis

Similar documents
Collaborators from SDM Center, CPES, GPSC, GSEP, Sandia, ORNL

Parallel, In Situ Indexing for Data-intensive Computing. Introduction

Adaptable IO System (ADIOS)

Canopus: Enabling Extreme-Scale Data Analytics on Big HPC Storage via Progressive Refactoring

EMPRESS Extensible Metadata PRovider for Extreme-scale Scientific Simulations

CSE6230 Fall Parallel I/O. Fang Zheng

Visualization and Analysis for Near-Real-Time Decision Making in Distributed Workflows

PROGRAMMING AND RUNTIME SUPPORT FOR ENABLING DATA-INTENSIVE COUPLED SCIENTIFIC SIMULATION WORKFLOWS

Big Data in Scientific Domains

Network Needs of US-China Fusion Research Collaborations

High-Performance Outlier Detection Algorithm for Finding Blob-Filaments in Plasma

Scibox: Online Sharing of Scientific Data via the Cloud

In situ data processing for extreme-scale computing

Introduction to High Performance Parallel I/O

The Fusion Distributed File System

AS big data has increasing influence on our daily life

SDS: A Framework for Scientific Data Services

PreDatA - Preparatory Data Analytics on Peta-Scale Machines

The National Fusion Collaboratory

IME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning

Using a Robust Metadata Management System to Accelerate Scientific Discovery at Extreme Scales

Improved Solutions for I/O Provisioning and Application Acceleration

Storage for HPC, HPDA and Machine Learning (ML)

Short Talk: System abstractions to facilitate data movement in supercomputers with deep memory and interconnect hierarchy

Big Data Challenges in Simulation-based Science

High Performance Data Analytics for Numerical Simulations. Bruno Raffin DataMove

...And eat it too: High read performance in write-optimized HPC I/O middleware file formats

Evaluation of Parallel I/O Performance and Energy with Frequency Scaling on Cray XC30 Suren Byna and Brian Austin

Scibox: Online Sharing of Scientific Data via the Cloud

On the Use of Burst Buffers for Accelerating Data-Intensive Scientific Workflows

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

Toward portable I/O performance by leveraging system abstractions of deep memory and interconnect hierarchies

Research in Middleware Systems For In-Situ Data Analytics and Instrument Data Analysis

XLDB /19/2011 Scott A. Klasky. To Exascale and beyond!

Percipient StorAGe for Exascale Data Centric Computing Computing for the Exascale

FES Science Network Requirements. Report of the Fusion Energy Sciences Network Requirements Workshop Conducted March 13 and 14, 2008

UK LUG 10 th July Lustre at Exascale. Eric Barton. CTO Whamcloud, Inc Whamcloud, Inc.

Tuning I/O Performance for Data Intensive Computing. Nicholas J. Wright. lbl.gov

Benchmarking the Performance of Scientific Applications with Irregular I/O at the Extreme Scale

Lustre architecture for Riccardo Veraldi for the LCLS IT Team

Building Simulation Modelers Are we big data ready? Jibonananda Sanyal and Joshua New Oak Ridge National Laboratory

Oh, Exascale! The effect of emerging architectures on scien1fic discovery. Kenneth Moreland, Sandia Na1onal Laboratories

Persistent Data Staging Services for Data Intensive In-situ Scientific Workflows

Isilon: Raising The Bar On Performance & Archive Use Cases. John Har Solutions Product Manager Unstructured Data Storage Team

N. Marusov, I. Semenov

High Throughput Data Movement

Microsoft Developer Day

Lustre* - Fast Forward to Exascale High Performance Data Division. Eric Barton 18th April, 2013

Multi-Resolution Streams of Big Scientific Data: Scaling Visualization Tools from Handheld Devices to In-Situ Processing

IBM Spectrum Scale IO performance

Skel: Generative Software for Producing Skeletal I/O Applications

Canopus: Enabling Extreme-Scale Data Analytics on Big HPC Storage via Progressive Refactoring

Toward a first-principles integrated simulation of tokamak edge plasmas

The Canadian CyberSKA Project

Preparing GPU-Accelerated Applications for the Summit Supercomputer

Modeling and Implementation of an Asynchronous Approach to Integrating HPC and Big Data Analysis

WHITE PAPER QUANTUM S XCELLIS SCALE-OUT NAS. Industry-leading IP Performance for 4K, 8K and Beyond

Fast Forward I/O & Storage

VisIt Libsim. An in-situ visualisation library

EXTREME SCALE DATA MANAGEMENT IN HIGH PERFORMANCE COMPUTING

SAP IQ Software16, Edge Edition. The Affordable High Performance Analytical Database Engine

Data Analytics and Storage System (DASS) Mixing POSIX and Hadoop Architectures. 13 November 2016

Co-existence: Can Big Data and Big Computation Co-exist on the Same Systems?

Store Process Analyze Collaborate Archive Cloud The HPC Storage Leader Invent Discover Compete

Data Movement & Storage Using the Data Capacitor Filesystem

Do You Know What Your I/O Is Doing? (and how to fix it?) William Gropp

Bridging the Gap Between High Quality and High Performance for HPC Visualization

Mahanaxar: Quality of Service Guarantees in High-Bandwidth, Real-Time Streaming Data Storage

Addressing the Variety Challenge to Ease the Systemization of Machine Learning in Earth Science

Data-Intensive Science Using GPUs. Alex Szalay, JHU

Extreme I/O Scaling with HDF5

The DEEP-ER take on I/O

High Speed Asynchronous Data Transfers on the Cray XT3

Approaching the Petabyte Analytic Database: What I learned

ScalaIOTrace: Scalable I/O Tracing and Analysis

Understanding Primary Storage Optimization Options Jered Floyd Permabit Technology Corp.

CA485 Ray Walshe Google File System

EMPRESS Extensible Metadata PRovider for Extreme-scale Scientific Simulations.

PHX: Memory Speed HPC I/O with NVM. Pradeep Fernando Sudarsun Kannan, Ada Gavrilovska, Karsten Schwan

Navigational Data Management. Joshua Stillerman, Martin Greenwald, John Wright MIT Plasma Science and Fusion Center

An Exploration into Object Storage for Exascale Supercomputers. Raghu Chandrasekar

Science-as-a-Service

Large Data Visualization

Analyzing the High Performance Parallel I/O on LRZ HPC systems. Sandra Méndez. HPC Group, LRZ. June 23, 2016

Data Movement & Tiering with DMF 7

XpressSpace: a programming framework for coupling partitioned global address space simulation codes

The Hopper System: How the Largest* XE6 in the World Went From Requirements to Reality! Katie Antypas, Tina Butler, and Jonathan Carter

icore: A GEOINT Processing Framework and Incubator

Lustre overview and roadmap to Exascale computing

Oncilla - a Managed GAS Runtime for Accelerating Data Warehousing Queries

ParaMEDIC: Parallel Metadata Environment for Distributed I/O and Computing

I/O at JSC. I/O Infrastructure Workloads, Use Case I/O System Usage and Performance SIONlib: Task-Local I/O. Wolfgang Frings

Structured Streams: Data Services for Petascale Science Environments

REPORT DOCUMENTATION PAGE

DDN Annual High Performance Computing Trends Survey Reveals Rising Deployment of Flash Tiers & Private/Hybrid Clouds vs.

The Data exacell DXC. J. Ray Scott DXC PI May 17, 2016

Realtime Data Analytics at NERSC

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

ENERGY-EFFICIENT VISUALIZATION PIPELINES A CASE STUDY IN CLIMATE SIMULATION

VISUALISATION AND ANALYSIS

Transcription:

Stream Processing for Remote Collaborative Data Analysis Scott Klasky 146, C. S. Chang 2, Jong Choi 1, Michael Churchill 2, Tahsin Kurc 51, Manish Parashar 3, Alex Sim 7, Matthew Wolf 14, John Wu 7 1 ORNL, 2 PPPL, 3 Rutgers, 4 GT, 5 SBU, 6 UTK, 7 LBNL STREAM 2016 Tysons, VA

Next Generation DOE computing File System and network bandwidth does not keep up with computing power

Big Data in Fusion Science: ITER example Volume: Initially 90 TB per day, 18 PB per year, maturing to 2.2 PB per day, 440 PB per year Value: All data are taken from expensive instruments for valuable reasons. Velocity: Peak 50 GB/s, with near real-time analysis needs Variety: ~100 different types of instruments and sensors, numbering in the thousands, producing interdependent data in various formats Veracity: The quality of the data can vary greatly depending upon the instruments and sensors. The pre-iter superconducting fusion experiments outside of US will also produce increasingly bigger data (KSTAR, EAST, Wendelstein 7-X, and JT60-SU later).

Streaming data from Simulations Whole Device Modeling the next fusion grand challenge From 1992 NSF grand challenge Now a possible Exascale Computing Application Much more difficult than original thought Couple Codes from different time scales, different physics, different spatial regimes. Very large data Edge simulation today producing about 100 PB in 10 days. Couple MHD effects, Core effects,. 1 EB in 2 days. Need data triage techniques Separate out information from data in near-real-time Very large number of observables Large number of workflows for each set of analysis Desire to understand what do we do on machine, off machine 4

Collaborative Nature of Science 100 s different diagnostics which produce data of different sizes/velocities Different scientist have workflows Fusion simulations produce 100 s of different variables Goal: run all analysis to make near-real-time decisions Realization: V s are too large: Prioritize which data gets processed when, where

Feature Extraction: Near Real Time Detection of Blobs Fusion Plasma blobs Lead to the loss of energy from tokamak plasmas Could damage multi-billion tokamak The experimental facility may not have enough computing power for the necessary data processing Distributed in transient processing Make more processing power available Allow more scientists to participate in the data analysis operations and monitor the experiment remotely Enable scientists to share knowledge and processes Blobs in fusion reaction (Source: EPSI project) Blob trajectory

ICEE, Enabling International Collaborations Example: KSTAR ECEI Sample Workflow Objective: To enable remote scientists to study ECE-Image movies of blobby turbulence and instabilities between experimental shots in near real-time. Input: Raw ECEi voltage data (~550MB/s, over 300 seconds in the future) + Metadata (experimental setting) Requirement: Data transfer, processing, and feedback within <15min (inter-shot time) Implementation: distributed data processing with ADIOS ICEE method Measure KSTAR Digitizer Buffer Server KSTAR Storage WAN Transfer (ADIOS) ADIOS (Staging) KISTI/ORNL Demo at SC14 Data post processing/ Making movies in Parallel Movie/physic files Feedback to next or next day shot Postech KSTAR (PPPL, GA, MIT, )

ICEE, Enabling Fusion Collaboration Data Fusion Objective: Enable comparisons of simulation (pre/post) and experiment at remote locations Input: Gas Puff Imaging (GPI) fast camera images from NSTX and XGC1 edge simulation data Output: Blob physics Requirement: Complete in near real-time for inter-shot experimental comparison, experiment-simulation validation or simulation monitoring Implementation: distributed data processing with ADIOS ICEE method, optimized detection algorithms for near real-time analysis Experiment (NSTX GPI) Simulation (XGC1)

ADIOS Abstraction Unifies Local And Remote I/O I/O Componentization for Data-at- Rest and Data-in-Motion Service Oriented Architecture for Extreme scaling computing Self Describing data movement/storage Main paper to cite Q. Liu, J. Logan, Y. Tian, H. Abbasi, N. Podhorszki, J. Choi, S. Klasky, R. Tchoua, J. Lofstead, R. Oldfield, M. Parashar, N. Samatova, K. Schwan, A. Shoshani, M. Wolf, K. Wu, W. Yu, Hello ADIOS: the challenges and lessons of developing leadership class I/O frameworks, Concurrency and Computation: Practice and Experience, 2013 Data API Core ADIOS components BPC BP container BPC spatial aggregation BPC I/O Aggregation BP Posix BP MPI-IO NetCDF4 GRIB2 Control API Buffering (generic) temporal aggregation compression HDF5 serial HDF5 parallel Staging Execution engine FastBit Indexing Data Services I/O Services User created output

The ADIOS-BP Stream/File format Ensemble of chunks = file All data chunks are from a single producer MPI process, Single diagnostic Ability to create a separate metadata file when sub-files are generated Allows variables to be individually compressed Has a schema to introspect the information Has workflows embedded into the data streams Format is for data-in-motion and data-at-rest

Auditing data during the streaming phase Streaming Data Stream Information Too much data to move in too little time Storage sizes/speed doesn t keep up with making NRT decisions Fit the data with a model From our physics understanding From our understanding of the entropy in the data From our understanding of the error in the data Change data into data = model + information (f = F + df) Streaming Information Reconstruct on-the-fly Query data from the remote source 11

I/O abstraction of data staging Normal 1 2 3 4 Application HPC System P F S Application Staging nodes P F S Application Analytics P F S Application Analy tics Viz P F S 5 Application HPC System Analytics hpc system Analy tics Viz ICEE F S 6 Diagnostic Analytics Experiment

Hybrid Staging Asynchronous Data Transfer S3D-Box Use compute and deep-memory hierarchies to optimize overall workflow for power vs. performance tradeoffs Abstract complex/deep memory hierarchy access Placement of analysis and visualization tasks in a complex system Impact of network data movement compared to memory movement Abstraction allows staging On-same core On different cores On different nodes On different machines Through the storage system ADIOS In situ Analysis and Visualization ADIOS ADIOS In transit Analysis Visualization ADIOS

ICEE System Development With ADIOS Data Generation Analysis Analysis FastBit Indexing Raw Data Index Data Source Site ICEE Server FastBit Query Data Flow Data Hub (Staging) Wide Area Network Feedback (WAN) Analysis Data Hub (Staging) Remote Client Sites Features ADIOS provides an overlay network to share data and give feedbacks Stream data processing supports stream-based IO to process pulse data In transit processing provides remote memory-to-memory mapping between data source (data generator) and client (data consumer) Indexing and querying with FastBit technology Analysis Analysis

Interactive Supercomputing from Singapore to Austin

Technology is also being used for SKA (J. Wang) The largest radio telescope in the world 1.5 billion project 11 member countries 2023-2030 Phase 2 constructed Currently conceptual design & preliminary benchmarks! Compute Challenge: 100 PFLOPS Data Challenge: ExaBytes per day Challenge is to run time-division correlator and then write output data to a parallel filesystem https://cug.org/proceedings/cug2014_proceedings/includes/files/pap121-file2.pdf