Warehouse- Scale Computing and the BDAS Stack

Similar documents
Spark, Shark and Spark Streaming Introduction

Shark. Hive on Spark. Cliff Engle, Antonio Lupher, Reynold Xin, Matei Zaharia, Michael Franklin, Ion Stoica, Scott Shenker

Apache Spark is a fast and general-purpose engine for large-scale data processing Spark aims at achieving the following goals in the Big data context

Resilient Distributed Datasets

UNIFY DATA AT MEMORY SPEED. Haoyuan (HY) Li, Alluxio Inc. VAULT Conference 2017

Hadoop 2.x Core: YARN, Tez, and Spark. Hortonworks Inc All Rights Reserved

Lecture 11 Hadoop & Spark

Database Architecture 2 & Storage. Instructor: Matei Zaharia cs245.stanford.edu

2/26/2017. Originally developed at the University of California - Berkeley's AMPLab

SparkStreaming. Large scale near- realtime stream processing. Tathagata Das (TD) UC Berkeley UC BERKELEY

The Datacenter Needs an Operating System

Page 1. Goals for Today" Background of Cloud Computing" Sources Driving Big Data" CS162 Operating Systems and Systems Programming Lecture 24

April Copyright 2013 Cloudera Inc. All rights reserved.

CS6453. Data-Intensive Systems: Rachit Agarwal. Technology trends, Emerging challenges & opportuni=es

Big Data Architect.

MapR Enterprise Hadoop

Fast, Interactive, Language-Integrated Cluster Computing

Computations with Bounded Errors and Response Times on Very Large Data

Spark & Spark SQL. High- Speed In- Memory Analytics over Hadoop and Hive Data. Instructor: Duen Horng (Polo) Chau

BIG DATA AND HADOOP ON THE ZFS STORAGE APPLIANCE

A BigData Tour HDFS, Ceph and MapReduce

Topics. Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples

DATA SCIENCE USING SPARK: AN INTRODUCTION

CIS 601 Graduate Seminar. Dr. Sunnie S. Chung Dhruv Patel ( ) Kalpesh Sharma ( )

Linear Regression Optimization

Big Data Infrastructures & Technologies

Announcements. Reading Material. Map Reduce. The Map-Reduce Framework 10/3/17. Big Data. CompSci 516: Database Systems

Big Data and Object Storage

EsgynDB Enterprise 2.0 Platform Reference Architecture

Spark. In- Memory Cluster Computing for Iterative and Interactive Applications

Apache Spark Graph Performance with Memory1. February Page 1 of 13

SCALABLE, LOW LATENCY MODEL SERVING AND MANAGEMENT WITH VELOX

IBM Data Science Experience White paper. SparkR. Transforming R into a tool for big data analytics

CompSci 516: Database Systems

Hadoop Beyond Batch: Real-time Workloads, SQL-on- Hadoop, and thevirtual EDW Headline Goes Here

Nowcasting. D B M G Data Base and Data Mining Group of Politecnico di Torino. Big Data: Hype or Hallelujah? Big data hype?

Summary of Big Data Frameworks Course 2015 Professor Sasu Tarkoma

Spark. In- Memory Cluster Computing for Iterative and Interactive Applications

A Platform for Fine-Grained Resource Sharing in the Data Center

Applied Spark. From Concepts to Bitcoin Analytics. Andrew F.

Introduction to Hadoop. Owen O Malley Yahoo!, Grid Team

Making the Most of Hadoop with Optimized Data Compression (and Boost Performance) Mark Cusack. Chief Architect RainStor

Trends and Challenges in Big Data

Cloud Computing & Visualization

MODERN BIG DATA DESIGN PATTERNS CASE DRIVEN DESINGS

Shark: SQL and Rich Analytics at Scale. Michael Xueyuan Han Ronny Hajoon Ko

Data Storage and Query Answering. Data Storage and Disk Structure (2)

Data Platforms and Pattern Mining

Today s content. Resilient Distributed Datasets(RDDs) Spark and its data model

Discretized Streams. An Efficient and Fault-Tolerant Model for Stream Processing on Large Clusters

SoftFlash: Programmable Storage in Future Data Centers Jae Do Researcher, Microsoft Research

Oracle Big Data Fundamentals Ed 1

Big Data Hadoop Stack

The Stratosphere Platform for Big Data Analytics

Introduction to MapReduce Algorithms and Analysis

Increasing Performance of Existing Oracle RAC up to 10X

HDInsight > Hadoop. October 12, 2017

MapReduce, Hadoop and Spark. Bompotas Agorakis

Table 1 The Elastic Stack use cases Use case Industry or vertical market Operational log analytics: Gain real-time operational insight, reduce Mean Ti

Massive Online Analysis - Storm,Spark

Spark. Cluster Computing with Working Sets. Matei Zaharia, Mosharaf Chowdhury, Michael Franklin, Scott Shenker, Ion Stoica.

CSE 451: Operating Systems Spring Module 12 Secondary Storage. Steve Gribble

Qunar Performs Real-Time Data Analytics up to 300x Faster with Alluxio

Things Every Oracle DBA Needs to Know about the Hadoop Ecosystem. Zohar Elkayam

a Spark in the cloud iterative and interactive cluster computing

How Apache Hadoop Complements Existing BI Systems. Dr. Amr Awadallah Founder, CTO Cloudera,

CSE 6242 / CX 4242 Course Review. Duen Horng (Polo) Chau Associate Director, MS Analytics Assistant Professor, CSE, College of Computing Georgia Tech

MISO: Souping Up Big Data Query Processing with a Multistore System

Lambda Architecture with Apache Spark

Webinar Series TMIP VISION

Spark: A Brief History.

Big Data Hadoop Developer Course Content. Big Data Hadoop Developer - The Complete Course Course Duration: 45 Hours

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Warehouse-Scale Computing

MapReduce Spark. Some slides are adapted from those of Jeff Dean and Matei Zaharia

Comparative Study of Apache Hadoop vs Spark

FlashGrid Software Enables Converged and Hyper-Converged Appliances for Oracle* RAC

Backtesting with Spark

Efficiency at Scale. Sanjeev Kumar Director of Engineering, Facebook

Sparrow. Distributed Low-Latency Spark Scheduling. Kay Ousterhout, Patrick Wendell, Matei Zaharia, Ion Stoica

An Introduction to Apache Spark

Programming Systems for Big Data

CSE 451: Operating Systems Spring Module 12 Secondary Storage

Cloud Computing 2. CSCI 4850/5850 High-Performance Computing Spring 2018

Achieving Horizontal Scalability. Alain Houf Sales Engineer

Big Data Programming: an Introduction. Spring 2015, X. Zhang Fordham Univ.

Principal Software Engineer Red Hat Emerging Technology June 24, 2015

Blended Learning Outline: Developer Training for Apache Spark and Hadoop (180404a)

Stages of Data Processing

Correlation based File Prefetching Approach for Hadoop

Bringing Data to Life

CS 61C: Great Ideas in Computer Architecture. MapReduce

Big Data and Cloud Computing

Approaching the Petabyte Analytic Database: What I learned

Ambry: LinkedIn s Scalable Geo- Distributed Object Store

Apache Spark 2.0. Matei

Hortonworks and The Internet of Things

Accelerating Hadoop Applications with the MapR Distribution Using Flash Storage and High-Speed Ethernet

Overview. Prerequisites. Course Outline. Course Outline :: Apache Spark Development::

The Impact of SSD Selection on SQL Server Performance. Solution Brief. Understanding the differences in NVMe and SATA SSD throughput

Storage Adapter Testing Report

Transcription:

Warehouse- Scale Computing and the BDAS Stack Ion Stoica UC Berkeley UC BERKELEY

Overview Workloads Hardware trends and implications in modern datacenters BDAS stack

What is Big Data used For? Reports, e.g.,» Track business processes, transactions Diagnosis, e.g.,» Why is user engagement dropping?» Why is the system slow?» Detect spam, worms, viruses, DDoS attacks Decisions, e.g.,» Decide what feature to add» Decide what ad to show» Block worms, viruses, Data is as useful as the decisions it enables

Data Processing Goals Low latency queries on historical data: enable faster decisions» E.g., identify why a site is slow and fix it Low latency queries on live data (streaming): enable decisions on real- time data» E.g., detect & block worms in real- time (a worm may infect 1mil hosts in 1.3sec) Sophisticated data processing: enable better decisions» E.g., anomaly detection, trend analysis

Typical Datacenter Node RAM (64-256GB) 8-16 cores Disk x10 (10-30TB) (10 GB/s) x 4 (50-100MB/s) x 10 CPU:» 8-16 cores» Transfer rate: quad channel DDR3: 4x10GB/s RAM: 64-256 GB Storage: 8-12 disks, each» Capacity: 1-3TB» Transfer rate: 50-100MB/s

Typical Datacenter Node 8-16 cores 0.5-1GB/s Disk x10 (10-30TB) Node 0.01-1GB/s 40 GB/s RAM (64-256GB) Network Top- of- rack switch Nodes in Nodes in same rack other rack

Typical Datacenter Node 0.8-2GB/s Nodes in Nodes in same rack other rack 0.5-1GB/s Disk x10 (10-30TB) Node 0.01-1GB/s 8-16 cores 40 GB/s RAM (64-256GB) Network Top- of- rack switch SSDs x4 (1-4TB) SSD drive: 256GB- 1TB 200-500 MB/s (SATA 3)

How Much Does $1,000 Buy You Today? 15-20TB disk storage (consumer grade disks) 1 TB of SSD storage (SATA- 3) 0.2TB of RAM

Memory Capacity Trends 0.8-2GB/s Nodes in Nodes in same rack other rack 0.2-1GB/s Disk x10 (10-30TB) Node 0.01-1GB/s 8-16 cores 40 GB/s RAM (64-256GB) Network Capacity doubles every 18 months (Moore s law) Top- of- rack switch SSDs x4 (1-4TB)

Disk Capacity Trends 0.8-2GB/s 0.2-1GB/s Disk x10 (10-30TB) Node 0.01-1GB/s 8-16 cores 40 GB/s RAM (64-256GB) Network Top- of- rack switch Nodes in Nodes in same rack other rack Capacity doubles every 18 months SSDs x5 (1-4TB)

SSD Capacity Trends Capacity doubles every 18 months (Moore s law) 0.8-2GB/s 40 GB/s Nodes in Nodes in same rack other rack 0.2-1GB/s Disk x10 (10-30TB) Node 0.01-1GB/s 8-16 cores RAM (64-256GB) Network Top- of- rack switch SSDs x5 (1-4TB)

CPU Trends Top- of- rack switch 0.8-2GB/s Nodes in Nodes in same rack other rack 0.2-1GB/s Disk x10 (10-30TB) Node 0.01-1GB/s 8-16 cores 40 GB/s RAM (64-256GB) Network Frequency no longer increasing # of cores doubles every 24 months SSDs x5 (1.5-5TB) source:http://www.drdobbs.com/parallel/design- for- manycore- systems/219200099

Network I/O Trends 40 GB/s 0.8-2GB/s Nodes in Nodes in same rack other rack 0.2-1GB/s Disk x10 (10-30TB) Node 0.01-1GB/s 8-16 cores RAM (64-256GB) Top- of- rack switch Network Server Network I/O doubling every 24 months SSDs x5 (1-4TB) source: http://www.ospmag.com/issue/article/time-isnot-always-on-our-side

Network I/O Trends 1-2.5GB/s Nodes in Nodes in same rack other rack 0.2-1GB/s Disk x10 (10-30TB) 0.01-1GB/s 8-16 cores 40 GB/s RAM (64-256GB) Top- of- rack switch Network Memory (i.e., DDR) bandwidth doubling every ~36 months SSDs x5 (1.5-5TB) Node source: http://www.theregister.co.uk/2009/05/26/rambus_pitches_xdr2/

SSD Throughput Trends Disk x10 (10-30TB) Node SSD (e.g., SATA) transfer rate doubles every 36 months 40 GB/s 0.8-2GB/s 0.2-1GB/s 0.01-1GB/s 8-16 cores Network RAM (64-256GB) Top- of- rack switch Bandwidth (Gbps) 100 10 SSDs x1 5 (1-4TB) Nodes in Nodes in same rack other rack 2003 2006 2009 2012

Disk Throughput Trends 0.8-2GB/s Nodes in Nodes in same rack other rack 0.2-1GB/s Disk x10 (10-30TB) 0.01-1GB/s 8-16 cores 40 GB/s RAM (64-256GB) Top- of- rack Disk bandwidth increases switch Network very slowly SSDs x5 (1.5-5TB) Node Source: Freitas et al, 2011

Trend Summary Rate doubles every 24 months Capacity doubles every 18 months Rate doubles every 36 months Rate doubles every 36 months Freq. no longer increasing; # of cores doubles every 24 months Rate increases very slowly Capacity doubles every 18 months Capacity doubles every 18 months

Trend Summary (cont d) Storage is cheap and capacity increases exponentially Most transfer rates increase exponentially but slower à gap between capacity and transfer rate increases Multiple channels/disks alleviate problem but don t come for free» E.g., disk striping increases block size Datacenter apps must carefully select where to place computations & data

Challenge and Opportunity Accessing disk very slow: 1,000s to read/write 100GB from/to disk» Transfer rate not increasing» Will get worse: 512 GB per node in one- two years» Faster to access remote memory!» SSDs not widely deployed in datacenters

Challenge and Opportunity (cont d) A few node cluster = 1TB RAM: enough to handle many large datasets» E.g., 1 billion users, 1KB metadata (on average) # of cores doubles only every 24 months à memory/core increases exponentially Judiciously using RAM is key

Existing Open Stack..mostly focused on large on- disk datasets» Massive data and sophisticated data processing, but slow We add RAM to the mix» Enable interactive queries and data streaming: speedup queries and iterative algorithms by up to 30x» Dramatically increase ability to explore / mine the data (SSDs in the future)

BDAS Software Stack Leverage open source ecosystem (e.g., HDFS, Hadoop) Abstractions to take advantage of storage hierarchy» Many real- world working sets fit in memory à RDDs» Controllable data placement to minimize communication Virtualize cluster resources (Mesos)» Allow multiple frameworks to share cluster and data à Resource offers Simplify parallel programming» Scala interface to Spark» Shark distributed SQL engine

Project History Mesos started in early 2009, open sourced 2010 Spark started in late 2009, open sourced 2010 Shark started summer 2011, alpha April 2012 Used at Twitter, Foursquare, Klout, Quantifind, Conviva, Yahoo! Research, Airbnb & others

Today s Open Analytics Stack Hadoop Stack (e.g., Hive, Pig, ) HDFS Other Frameworks (Storm, MPI, ) Frmwk. Storage

BDAS Software Stack Hadoop Stack (e.g., Hive, Pig, ) Other Frameworks (Storm, MPI) Frmwk. Mesos Management platform that allows multiple framework to share cluster HDFS Resource Mgmt. Storage

BDAS Software Stack Spark In- memory processing framework: Hadoop Iterative computations Other Stack Interactive (e.g., queries Frameworks Hive, Pig, ) (Storm, MPI) Mesos HDFS Frmwk. Resource Mgmt. Storage

BDAS Software Stack HQL (Hive) Interface Shark Spark HIVE on top of Spark Interactive queries for Hadoop data fitting in memory Other Stack (e.g., Frameworks Hive, Pig, ) (Storm, MPI) Mesos HDFS Frmwk. Resource Mgmt. Storage

BDAS Software Stack Shark SparkStreaming Hadoop Stack (e.g., Hive, Pig, ) Low- latency Mesos streaming processing for very large data HDFS Other Frameworks (Storm, MPI) Frmwk. Resource Mgmt. Storage

BDAS Software Stack BlinkDB Shark SparkStreaming Provide bounded error and bounded response Hadoop time on Other Stack very large (e.g., data Frameworks Hive, Pig, ) (Storm, MPI) Mesos HDFS Frmwk. Resource Mgmt. Storage

BDAS Software Stack BlinkDB Shark SparkStreaming Hadoop Stack (e.g., Hive, Pig, ) Other Frameworks (Storm, MPI) Frmwk. Trex Share in- memory Mesos storage (RDDs) across multiple frameworks HDFS Data Resource Mgmt. Storage

Summary Today: first iteration of BDAS Software Stack» Mesos: enable multiple frameworks to share cluster resources and data» Spark: enable interactive and iterative computations through the use of RDDs» Shark: enable interactive Hive queries Next: full stack to allow users to trade between answer (1) quality, (2) response time, and (3) cost Fully compatible with open standards