China Big Data and HPC Initiatives Overview. Xuanhua Shi

Similar documents
Gearing Hadoop towards HPC sytems

Optimizing Apache Spark with Memory1. July Page 1 of 14

Apache Spark is a fast and general-purpose engine for large-scale data processing Spark aims at achieving the following goals in the Big data context

Practical Near-Data Processing for In-Memory Analytics Frameworks

Accelerate Big Data Insights

Challenges for Data Driven Systems

The Role of Database Aware Flash Technologies in Accelerating Mission- Critical Databases

Distributed File Systems II

Using Alluxio to Improve the Performance and Consistency of HDFS Clusters

Processing of big data with Apache Spark

SOLVING THE DRAM SCALING CHALLENGE: RETHINKING THE INTERFACE BETWEEN CIRCUITS, ARCHITECTURE, AND SYSTEMS

Innovator, Disruptor or Laggard, Where will your storage applications live? Next generation storage

The Fusion Distributed File System

Accelerating Spark Workloads using GPUs

PHX: Memory Speed HPC I/O with NVM. Pradeep Fernando Sudarsun Kannan, Ada Gavrilovska, Karsten Schwan

Efficient Data Mapping and Buffering Techniques for Multi-Level Cell Phase-Change Memories

CompSci 516: Database Systems

MapReduce: Simplified Data Processing on Large Clusters 유연일민철기

Oracle Exadata: Strategy and Roadmap

Scalable Web Programming. CS193S - Jan Jannink - 2/25/10

IQ for DNA. Interactive Query for Dynamic Network Analytics. Haoyu Song. HUAWEI TECHNOLOGIES Co., Ltd.

Cloud Computing & Visualization

Announcements. Reading Material. Map Reduce. The Map-Reduce Framework 10/3/17. Big Data. CompSci 516: Database Systems

Leveraging Software-Defined Storage to Meet Today and Tomorrow s Infrastructure Demands

2/26/2017. Originally developed at the University of California - Berkeley's AMPLab

Next-Generation Cloud Platform

Shark. Hive on Spark. Cliff Engle, Antonio Lupher, Reynold Xin, Matei Zaharia, Michael Franklin, Ion Stoica, Scott Shenker

PLATFORM AND SOFTWARE AS A SERVICE THE MAPREDUCE PROGRAMMING MODEL AND IMPLEMENTATIONS

Flash Storage Complementing a Data Lake for Real-Time Insight

Chapter 4: Apache Spark

Huge market -- essentially all high performance databases work this way

Data Processing at the Speed of 100 Gbps using Apache Crail. Patrick Stuedi IBM Research

SpongeFiles: Mitigating Data Skew in MapReduce Using Distributed Memory. Khaled Elmeleegy Turn Inc.

Workload Characterization and Optimization of TPC-H Queries on Apache Spark

Linear Regression Optimization

Research challenges in data-intensive computing The Stratosphere Project Apache Flink

Lecture 9: MIMD Architectures

Fusion iomemory PCIe Solutions from SanDisk and Sqrll make Accumulo Hypersonic

IBM Data Science Experience White paper. SparkR. Transforming R into a tool for big data analytics


Spark Overview. Professor Sasu Tarkoma.

Planning For Persistent Memory In The Data Center. Sarah Jelinek/Intel Corporation

Write a technical report Present your results Write a workshop/conference paper (optional) Could be a real system, simulation and/or theoretical

IBM Education Assistance for z/os V2R2

: A new version of Supercomputing or life after the end of the Moore s Law

Principal Software Engineer Red Hat Emerging Technology June 24, 2015

RAMCloud: Scalable High-Performance Storage Entirely in DRAM John Ousterhout Stanford University

NVMFS: A New File System Designed Specifically to Take Advantage of Nonvolatile Memory

Staged Memory Scheduling

Designing High-Performance Non-Volatile Memory-aware RDMA Communication Protocols for Big Data Processing

Middleware and Flash Translation Layer Co-Design for the Performance Boost of Solid-State Drives

Big Data Systems on Future Hardware. Bingsheng He NUS Computing

HDFS: Hadoop Distributed File System. CIS 612 Sunnie Chung

The Analysis Research of Hierarchical Storage System Based on Hadoop Framework Yan LIU 1, a, Tianjian ZHENG 1, Mingjiang LI 1, Jinpeng YUAN 1

HiTune. Dataflow-Based Performance Analysis for Big Data Cloud

Survey Paper on Traditional Hadoop and Pipelined Map Reduce

Row Buffer Locality Aware Caching Policies for Hybrid Memories. HanBin Yoon Justin Meza Rachata Ausavarungnirun Rachael Harding Onur Mutlu

A BigData Tour HDFS, Ceph and MapReduce

Chapter 5. The MapReduce Programming Model and Implementation

BigData and Map Reduce VITMAC03

IME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning

Strata: A Cross Media File System. Youngjin Kwon, Henrique Fingler, Tyler Hunt, Simon Peter, Emmett Witchel, Thomas Anderson

TITLE: PRE-REQUISITE THEORY. 1. Introduction to Hadoop. 2. Cluster. Implement sort algorithm and run it using HADOOP

Computer Architecture

A Review Paper on Big data & Hadoop

GridGain and Apache Ignite In-Memory Performance with Durability of Disk

Introduction to Hadoop and MapReduce

Diffusing Your Mobile Apps: Extending In-Network Function Virtualisation to Mobile Function Offloading

Hybrid Memory Platform

Massive Online Analysis - Storm,Spark

Bringing Data to Life

Konstantin Shvachko, Hairong Kuang, Sanjay Radia, Robert Chansler Yahoo! Sunnyvale, California USA {Shv, Hairong, SRadia,

PRESENTATION TITLE GOES HERE

Datacenter replication solution with quasardb

Strategic Briefing Paper Big Data

Enabling Cost-effective Data Processing with Smart SSD

Voldemort. Smruti R. Sarangi. Department of Computer Science Indian Institute of Technology New Delhi, India. Overview Design Evaluation

Topics. Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples

Software and Tools for HPE s The Machine Project

Bigtable: A Distributed Storage System for Structured Data. Andrew Hon, Phyllis Lau, Justin Ng

Diagnostics in Testing and Performance Engineering

Silent Shredder: Zero-Cost Shredding For Secure Non-Volatile Main Memory Controllers

Accelerating Enterprise Search with Fusion iomemory PCIe Application Accelerators

STORAGE LATENCY x. RAMAC 350 (600 ms) NAND SSD (60 us)

Distributed Systems. 22. Spark. Paul Krzyzanowski. Rutgers University. Fall 2016

Analytics in Spark. Yanlei Diao Tim Hunter. Slides Courtesy of Ion Stoica, Matei Zaharia and Brooke Wenig

Software-Defined Data Infrastructure Essentials

Fluentd + MongoDB + Spark = Awesome Sauce

Improved Solutions for I/O Provisioning and Application Acceleration

Multilevel Memories. Joel Emer Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology

Improving DRAM Performance by Parallelizing Refreshes with Accesses

Challenges in Data Stream Processing

Qunar Performs Real-Time Data Analytics up to 300x Faster with Alluxio

Clash of the Titans: MapReduce vs. Spark for Large Scale Data Analytics

IBM System Storage DCS3700

Sub-millisecond Stateful Stream Querying over Fast-evolving Linked Data

Hierarchy of knowledge BIG DATA 9/7/2017. Architecture

Introduction to Big-Data

CSCI-GA Database Systems Lecture 8: Physical Schema: Storage

Reducing DRAM Latency at Low Cost by Exploiting Heterogeneity. Donghyuk Lee Carnegie Mellon University

Transcription:

China Big Data and HPC Initiatives Overview Xuanhua Shi Services Computing Technology and System Laboratory Big Data Technology and System Laboratory Cluster and Grid Computing Laboratory Huazhong University of Science and Technology, Wuhan, China xhshi@hust.edu.cn

Outline n In-Memory Computing n New Funding by MOST n HPC Initiatives (2016-2020) n Big Data Initiatives (2016-2020)

In-Memory Computing: Lifting the Burden of Big Data Aberdeen Group 3

In-memory Database Traditional Computing in-memory Computing Database of Record Database of Record App. Data Application Code App. Data Application Code Persistency Recovery Post-processing Backup App. Data Application Code App. Data Application Code Main Memory (DRAM) Main Memory (DRAM) ØTraditional systems exchange data pages between memory and disk when computing with big data : Expensive IO cost Ø Speed of Disk: ms Speed of Memory: ns Ø In-Memory Computing: CPU reads data from memory and provides realtime data processing.

Downsides of DRAM Refresh Energy consumption: Each refresh consumes energy Performance degradation: DRAM bank unavailable while refreshed QoS/predictability impact: (Long) pause times during refresh Refresh rate limits DRAM capacity scaling

Emerging Non-volatile Memory (NVM) Technologies 6/17/16

Storage Class Memory SCM is a new class of data storage and memory devices SCM blurs the distinction between MEMORY (= fast, expensive, volatile ) and STORAGE (= slow, cheap, non-volatile) Characteristics of SCM Solid state, no moving parts Short access times (~ DRAM like, within an order-of-magnitude) Low cost per bit (DISK like, within an order-of-magnitude) Non-volatile ( ~ 10 years)

Reconstruction of Virtual Memory Architecture: Break the I/O Bottleneck 8

New In-Memory Computing Architecture Build DRAM + SCM hybrid hierarchical/parallel memory structure Memory Channel I/O Channel Disk Memory Channel Data I/O Channel New In-memory Computing Architecture The new in-memory computing architecture supports the shift from computing-centric to combination of computing and data 9

Challenges of Hybrid Memory Systems How should SCM-based (main) memory be organized? Partitioning Should DRAM be a cache or main memory, or configurable? What fraction? How many controllers?

Challenges of Hybrid Memory Systems Data allocation/movement (energy, performance, lifetime) Who manages allocation/movement? What are good control algorithms? How do we prevent degradation of service due to wearout? Design of cache hierarchy, memory controllers, OS Mitigate PCM shortcomings, exploit PCM advantages Persistent data can be randomly and synchronously addressed Huge non-volatile address spaces, memory-mapped DB, persistent objects Should SCM be used like I/O or like memory or in a totally new way?

Challenges of Hybrid Memory Systems Software Architecture Should one make SCM visible to applications software? If visible, in which form? New APIs, libraries, memory models, new I/O devices,... Databases, Business Intelligence and Streams are first impacted Data-intensive HPC - predictable execution time of complex business analytics - streaming search

Challenges of Enabling and Exploiting NVM Enabling NVM and hybrid memory How to tolerate errors? How to enable secure operation? How to tolerate performance and power shortcomings? How to minimize cost? Exploiting emerging technologies How to exploit non-volatility? How to minimize energy consumption? How to exploit NVM on chip?

Technology and System of In-Memory Computing for Big Data Processing Hybrid Memory Architecture for In-Memory Computing System System Software for In-Memory Computing System Parallel Processing for In-Memory Computing System Data Management for In-Memory Computing System 14

Technology and System of In-Memory Computing for Big Data Processing Project Overview Total budget: 170M RMB Period: January 2015 December 2017 Participants Inspur Huawei Sugon Shanghai Jiaotong University Huazhong University of Science and Technology Chongqing University National University of Defense Technology Huadong Normal University Jiangnan Computing Institute 15

Technology and System of In-Memory Computing for Big Data Processing Project Mission Hybrid NVM-based high reliable, massive storage, and low power in-memory computing system, the capacity of NVM in each node should be in TB level, supporting zero bootup System software and simulation platform for inmemory computing system Parallel processing system for in-memory computing system In-memory database for hybrid memory architecture to support decision making and other data management applications 16

System Architecture

Full System Simulator Applications Core Core Core Core Latency-Aware Shared Cache Data Migration Service Channel Redirect Mapping PCM Libs OS Migrator DRAM Software Level Migration Libs Hardware Level Migration Strategies Designed Based On: MARSSX86+ NVMain More flexible (supports NVMain DRAMSim HybridSim as main memory simulator) Simulate memory system precisely, easy to configure

Pros and Cons of DDPS User Program UDT UDF Distributed Data Process System Platform with memory management Stage1 StageN Produce data objects with different lifetime Apply Memory Y allocate N GC Benefit from the high-level objectoriented language The automatic memory management of platform effect the performance seriously

Deca: Exploiting Raw Data of In-Memory Data Objects in Distributed Data-Parallel Systems l Completely decomposing inmemory data objects to eliminate the references invocation and reduce frequent garbage collection in JVM l A system to automatic converse the user codes, decompose data objects and manage in-memory raw data l Speedup from 22.7x to 41.6x, compare with Apache Spark

Hadoop Engine: IO-intensive Map Task 1.I/O operations input split Hdfs collect Map buffer 2.I/O operations Local Disk Jvm sortandspill Merge-Sort (Combine) I/O wait file0 file1 file2 final merged file 3.I/O operations Shuffle NetWork final output 7.I/O operations Reduce Task 6.I/O operations Reduce file0 file1 file2 Merge 5.I/O operations Pull Sort Jvm 4.I/O operations 16/6/17 21

Mammoth: Memory-Centric MapReduce System A novel rule-based heuristic to prioritize memory allocation and revocation mechanism A multi-threaded execution engine, which realize global memory management Compatible with Hadoop Sources available at Github and ASF IEEE Computer Spotlight Execution Engine 10 22

Landscape of Disk-Based and In-Memory Data Management Systems (2014) 23

Outline n In-Memory Computing n New Funding by MOST n HPC Initiatives (2016-2020) n Big Data Initiatives (2016-2020)

HPC Initiatives (2016-2020) Two 100PFlops Supercomputer One is located Wuxi, another is located in Guangzhou E-scale architecture E-scale processors High-speed network HPC software stack Co-design: Aircraft design and Weather forecasting Some typical applications 25

Big Data Initiatives (2016-2020) Big data infrastructure New storage system Data flow based data analyzed stack Domain specific data management system Data-driven software Data analyze applications and Human-like intelligence From data to knowledge Large scale objects recognition 26

Thanks!