Canopus: Enabling Extreme-Scale Data Analytics on Big HPC Storage via Progressive Refactoring

Similar documents
Canopus: Enabling Extreme-Scale Data Analytics on Big HPC Storage via Progressive Refactoring

Compressing the Incompressible with ISABELA: In-situ Reduction of Spatio-Temporal Data

Fixed-PSNR Lossy Compression for Scientific Data

Compressing Intermediate Keys between Mappers and Reducers in SciHadoop

Stream Processing for Remote Collaborative Data Analysis

ISABELA for effective in situ compression of scientific data

The Fusion Distributed File System

EMPRESS Extensible Metadata PRovider for Extreme-scale Scientific Simulations

Analyzing Data Properties using Statistical Sampling Illustrated on Scientific File Formats

SZ-1.4: Significantly Improving Lossy Compression for Scientific Data Sets Based on Multidimensional Prediction and Error- Controlled Quantization

High-Performance Outlier Detection Algorithm for Finding Blob-Filaments in Plasma

Using a Robust Metadata Management System to Accelerate Scientific Discovery at Extreme Scales

Scibox: Online Sharing of Scientific Data via the Cloud

CSE6230 Fall Parallel I/O. Fang Zheng

ALACRITY: Analytics-Driven Lossless Data Compression for Rapid In-Situ Indexing, Storing, and Querying

Analyzing data properties using statistical sampling techniques illustrated on scientific file formats and compression features

Parallel Geospatial Data Management for Multi-Scale Environmental Data Analysis on GPUs DOE Visiting Faculty Program Project Report

Caching and Buffering in HDF5

Scalable Compression and Transmission of Large, Three- Dimensional Materials Microstructures

Efficient Lossy Compression for Scientific Data based on Pointwise Relative Error Bound

PROGRAMMING AND RUNTIME SUPPORT FOR ENABLING DATA-INTENSIVE COUPLED SCIENTIFIC SIMULATION WORKFLOWS

Collaborators from SDM Center, CPES, GPSC, GSEP, Sandia, ORNL

Analytics-driven Lossless Data Compression for Rapid In-situ Indexing, Storing, and Querying

Parallel, In Situ Indexing for Data-intensive Computing. Introduction

Simple and Powerful Animation Compression. Nicholas Fréchette Programming Consultant for Eidos Montreal

ScalaIOTrace: Scalable I/O Tracing and Analysis

IME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning

MIGRATORY COMPRESSION Coarse-grained Data Reordering to Improve Compressibility

Using Arithmetic Coding for Reduction of Resulting Simulation Data Size on Massively Parallel GPGPUs

Haar Wavelet Image Compression

Visualization and Analysis for Near-Real-Time Decision Making in Distributed Workflows

FastForward I/O and Storage: ACG 8.6 Demonstration

3 Data Storage 3.1. Foundations of Computer Science Cengage Learning

Z-checker: A framework for assessing lossy compression of scientific data

An Evolutionary Path to Object Storage Access

Data Compression. The Encoder and PCA

ENERGY-EFFICIENT VISUALIZATION PIPELINES A CASE STUDY IN CLIMATE SIMULATION

MSST 08 Tutorial: Data-Intensive Scalable Computing for Science

High Performance Data Analytics for Numerical Simulations. Bruno Raffin DataMove

Compressing CESM Data while Preserving Information

Compression of 3-Dimensional Medical Image Data Using Part 2 of JPEG 2000

For Mac and iphone. James McCartney Core Audio Engineer. Eric Allamanche Core Audio Engineer

A Breakthrough in Non-Volatile Memory Technology FUJITSU LIMITED

Fast Forward Storage & I/O. Jeff Layton (Eric Barton)

Digital Image Representation Image Compression

Novel Lossy Compression Algorithms with Stacked Autoencoders

These notes are designed to provide an introductory-level knowledge appropriate to understanding the basics of digital data formats.

Exponential Numbers ID1050 Quantitative & Qualitative Reasoning

An Adaptive Prediction-based Approach to Lossless Compression of Floating-point Volume Data

Getting it Right: Testing Storage Arrays The Way They ll be Used

Introduction to HPC Parallel I/O

CS 335 Graphics and Multimedia. Image Compression

Interactive Progressive Encoding System For Transmission of Complex Images

TrajStore: an Adaptive Storage System for Very Large Trajectory Data Sets

Adaptable IO System (ADIOS)

Big Data in Scientific Domains

Revealing Applications Access Pattern in Collective I/O for Cache Management

Hardware-Compatible Vertex Compression Using Quantization and Simplification

XLDB /19/2011 Scott A. Klasky. To Exascale and beyond!

Big Data in HPC. John Shalf Lawrence Berkeley National Laboratory

Topic 5 Image Compression

JULEA: A Flexible Storage Framework for HPC

Compression in Open Source Databases. Peter Zaitsev April 20, 2016

Data Reduction Techniques for Scientific Visualization and Data Analysis

Lecture 12: Compression

Design Tradeoffs for Data Deduplication Performance in Backup Workloads

Scibox: Online Sharing of Scientific Data via the Cloud

The State and Needs of IO Performance Tools

Distance Field Compression

VAPOR Product Roadmap. Visualization and Analysis Software Team October 2017

Unified VLSI Systolic Array Design for LZ Data Compression

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

TrajStore: an Adaptive Storage System for Very Large Trajectory Data Sets

Visual Analytics Tools for the Global Change Assessment Model. Michael Steptoe, Ross Maciejewski, & Robert Link Arizona State University

Improved Solutions for I/O Provisioning and Application Acceleration

Light Fields. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen

Data Transformation and Migration in Polystores

Improving Performance of Iterative Methods by Lossy Checkponting

Design, Implementation and Evaluation of a Task-parallel JPEG Decoder for the Libjpeg-turbo Library

COMP 4801 Final Year Project. Ray Tracing for Computer Graphics. Final Project Report FYP Runjing Liu. Advised by. Dr. L.Y.

On the Use of Burst Buffers for Accelerating Data-Intensive Scientific Workflows

Data Representation and Architecture Modeling Year 1 Exam preparation

DATABASE COMPRESSION. Pooja Nilangekar [ ] Rohit Agrawal [ ] : Advanced Database Systems

Large Data Visualization

Parallelizing Inline Data Reduction Operations for Primary Storage Systems

Exploratory Visualization of Petascale Particle Data in Nvidia DGX-1

7.5 Dictionary-based Coding

FPC: A High-Speed Compressor for Double-Precision Floating-Point Data

Viscous Fingers: A topological Visual Analytic Approach

Image Compression Algorithm and JPEG Standard

Lossy Compression of Scientific Data with Wavelet Transforms

Fast Forward I/O & Storage

HDF5 I/O Performance. HDF and HDF-EOS Workshop VI December 5, 2002

CSEP 521 Applied Algorithms Spring Lossy Image Compression

High Throughput Compression of Double-Precision Floating-Point Data

Understanding Primary Storage Optimization Options Jered Floyd Permabit Technology Corp.

NERSC Site Update. National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory. Richard Gerber

EFFICIENT DEISGN OF LOW AREA BASED H.264 COMPRESSOR AND DECOMPRESSOR WITH H.264 INTEGER TRANSFORM

Parallel I/O Libraries and Techniques

libhio: Optimizing IO on Cray XC Systems With DataWarp

Transcription:

Canopus: Enabling Extreme-Scale Data Analytics on Big HPC Storage via Progressive Refactoring Tao Lu*, Eric Suchyta, Jong Choi, Norbert Podhorszki, and Scott Klasky, Qing Liu *, Dave Pugmire and Matt Wolf, and Mark Ainsworth + * New Jersey Institute of Technology Oak Ridge National Laboratory + Brown University

Overview Background and related work Progressive data refactoring Conclusion 2

HPC systems Mission: illuminate phenomena that are often impossible to study in a laboratory [Oak Ridge National Lab, 2013] Climate impacts of energy use Fusion in a reactor not yet built Galaxy formation Methodology: Modeling and simulation, along with data exploration Data generation Storage Analysis Visualization 3

HPC systems (cont d) The big data management challenge [Shalf et. al., 2014] Worsening I/O bottleneck for exascale systems Exponentially increasing multi-scale, multi-physics data 4

Data compression in exascale (nextgeneration HPC) systems Goal: 10x to 100x data reduction ratio [Ian Foster et. al., 2017] Reduce data by at least 90% Data features Temporal and spatial High-dimensional Floating-point 5

Data compressors Lossless compression Deduplication GZIP FPC [Burtscher et. al., 2009] Lossy compression ZFP [Lindstrom et. al., 2014] ISABELA [Lakshminarasimhan et.al., 2011] SZ [Shen et. al., 2016] 6

Lossy compression Original Data 1. Array linearization 2. Data transformation 3. Curve fitting 4. Optimize the unpredictable Compressed data Workflow of curve fitting based compression (e.g. ISABELA and SZ) Floating-point Block: 0.368178-1.269298-0.904911-0.242216 1. Align to a common exponent Mantissa and exponent : 0.184089 * 2 1-0.684649 * 2 1-0.452456 * 2 1-0.121108 * 2 1 2a. Encode exponent 2b. Convert mantissa to integer Compressed low bits Integers: 848960939002912256-3157387122695243776 -2086581739634989568-558511819984848000 3. Orthogonal transform, reorder, and convert to unsigned integer Integers: 3733945919058534944 395516959657970744 4457920553677607264 580309805578545072 Compressed 4. Encode high bits coefficients Workflow of quantization and transformation based compression (e.g. ZFP) 7

Can floating-point compressors achieve a near 100x compression ratio? 8

Performance of compressors Compression ratio 400 300 200 100 0 fpc gzip isb zfp sz 30 20 10 0 Compression ratio 400 300 200 100 0 fpc gzip isb zfp sz 30 20 10 0 Data set Data set Relative error bound 0.000001 Relative error bound 0.001 9

Can floating-point compressors achieve a near 100x compression ratio? Yes. If Dataset contains a lot of identical values; Or, data values highly skew with moderate compression error bounds. Can floating-point compressors achieve a 10x compression ratio? Yes. For a lot of datasets with moderate compression error bounds. No. For most datasets. 10

What if the compression ratio is rushed to 100x by loosening error bounds? 11

Visualization and blob detection on compressed Dpot data.

Limitations of data compression by reducing floating-point accuracy Near 100x compression ratio is hardly achievable Lost data accuracy cannot be restored 13

Overview Background and related work Progressive data refactoring Conclusion 14

We propose Canopus Compressing HPC data in another dimension (resolution) Enabling progressive data refactoring User transparent implementation 15

Canopus I/O configuration

Canopus: basic idea Refactor the simulation results (via decimation) into a base dataset along with a series of deltas Base dataset is saved in fast devices, deltas in slow devices Base dataset can be used separately (at a lower resolution) for analysis Selected subset of deltas to be retrieved to restore data to a target accuracy MPI Simulation ADIOS Write API Canopus (I/O, refactoring, compression, placement, retrieval, restoration ) MPI_AGGREGATE ADIOS Kernel (buffering, metadata, scheduling, etc.) POSIX I/O Transport Dataspaces Node-local Storage (NVRAM, SSDs) Burst Buffer Remote Parallel File System Campaign Storage Storage Tiers Canopus in HPC Systems Data Analytics ADIOS Query API FLEXPATH MPI_LUSTRE 17

Canopus workflow HPC Simulations (high accuracy) Refactoring (decimation, compression) Retrieving & Reconstruction Base ST 1 Delta 2x ST 2 Analytics Pipeline 1 (low accuracy) base = L 4x Analytics Pipeline 2 (medium accuracy) base + delta 2x Delta full ST 3 Analytics Pipeline n (high accuracy) Storage Hierarchy base + delta 2x + delta full 18

Data refactoring Original Decimation Delta calculation 1. Mesh decimation Mesh (Full) Mesh (4x reduction) delta 2x 2. Delta calculation 3. Floating-point compression Full L4x delta full 19

Mesh decimation V l j V l i V l+1 i Delta Calculation For mesh data, it s common that each vertex corresponds to a value (floating-point) After triangular mesh decimation: V l h V l+1 h V l k V l+1 k V l+1 i = ½ (V l i + V l j) delta l n = F(V l n) - 1/3*F(V l+1 i) - 1/3*F(V l+1 j) - 1/3*F(V l+1 k) 20

Compression The floating-point values corresponding to vertexes are compressed using ZFP compressor A potential optimization to our framework is supporting adaptive compressors based on dataset features 21

Progressive data exploration (reverse the data refactoring procedures) I/O (read the base dataset and deltas) Decompression Restoration 22

Performance gain of Canopus for data analytics 23

Impact on Data Analytics Original 2x reduction 4x reduction 8x reduction 16x reduction 32x reduction 24

A quantitative evaluation of blob detection 25

Overview Storage stacks of HPC systems Progressive data refactoring Conclusion and future work 26

Conclusion Lossy compression may devastate the usefulness of data to achieve high compression ratio (such as 100x) It is critical to compress data in multiple orthogonal dimensions such as accuracy and resolution Canopus combines mesh compression and floatingpoint compression, possibly delivering a high compression ratio without devastate the usefulness of data 27

Future work Investigate the impact of lossy compression on analytical applications other than visualization Original data A == B, compressed data A == B? F(D) == F(D )? F is a function 28

References Oak Ridge National Lab, Solving Big Problems: Science and Technology at Oak Ridge National Laboratory, 2013 Foster, I., Ainsworth, M., Allen, B., Bessac, J., Cappello, F., Choi, J. Y., Yoo, S. (2017). Computing Just What You Need : Online Data Analysis and Reduction at Extreme Scales, 1 16. Shalf, J., Dosanjh, S., & Morrison, J. (2014). Top ten exascale research challenges, 1 25. Retrieved from http://link.springer.com/chapter/10.1007/978-3-642-19328-6_1 Burtscher, M., & Ratanaworabhan, P. (2009). FPC: A high-speed compressor for doubleprecision floating-point data. IEEE Transactions on Computers, 58(1), 18 31. Lindstrom, P. (2014). Fixed-rate compressed floating-point arrays. IEEE Transactions on Visualization and Computer Graphics, 20(12), 2674 2683. Lakshminarasimhan, S., Shah, N., Ethier, S., Klasky, S., Latham, R., Ross, R., & Samatova, N. F. (n.d.). Compressing the Incompressible with ISABELA : In-situ Reduction of Spatio- Temporal Data, 1 14. 29

Thanks & Questions 30