Scalable Table Stores: Tools for Understanding Advanced Key-Value Systems for Hadoop
|
|
- Logan Hood
- 5 years ago
- Views:
Transcription
1 Scalable Table Stores: Tools for Understanding Advanced Key-Value Systems for Hadoop Garth Gibson Professor, Carnegie Mellon Univ., & CTO, Panasas Inc. with Julio Lopez, Swapnil Patil, Milo Polte, Kai Ren, Wittawat Tantisiriroj, Lin Xiao, CMU to appear in SoCC October 2011 with Wittawat Tantisiriroj, Swapnil Patil, CMU Seung Woo Son, Sam Lang, Rob Ross, Argonne Nat Lab to appear in SC11 November 2011
2 The Future is Data-Led Expert human translator BLEU Score Usable translation Human-edittable translation Topic identification Useless Google ISI IBM+CMU UMD JHU+CU Edinburgh Systran Mitre NIST: translate 100 articles Arabic-English competition 2005 outcome: Google wins! Qualitatively better on 1st entry Brute force statistics with more data & compute!! FSC IEEE Intelligent Systems, March/April M words from UN translations 1 billion words of English grammar 1000 processor cluster 2
3 Science of Many Types is Data-Led Contact Field Comments J Lopez, CSD Astrophysics SDSS digital sky survey including spectroscopy, 50TB T Di Matteo, Physics Astrophysics Bigben BHCosmo hydrodynamics (1B particles simulated), 30TB F Gilman, Physics Astrophysics Large Synoptic Survey Telescope, LSST (2012) digital sky survey, 15TB/day C Langmead,CSD Biology Xray, NMR, CryoEM images; Sim d molecular dynamics trajectories J Bielak, CE Earth sciences USGS sensor images; Sim d 4D earthquake wavefields >10TB/run D Brumley, ECE Cyber security Worldwide Malware Archive; 2TB doubling each year O Mutlu, ECE Genomics 50GB per compressed genome sequencing; expands to TBs to process B Yu, ECE Neuroscience Neural recordings (electrodes, optical) for prosthetics; GB each J Callan, LTI Info Retrieval ClueWeb09, 25TB, 1B high rank web pages, 10 languages T Mitchell, MLD Machine Learning English sentences of ClueWeb for continuous automated reading (5TB) M Herbert, RI Image Understanding Flickr archive (>4TB); broadcast TV archive; street video; soldier video Y Sheikh, RI Virtual Reality Terascale VR sensor, 1000 camera+ 200 microphone, up to 5TB/sec C Guestrin, CSD Machine Learning Blog update archives, 2TB now + 2.7TB/yr (about 500K blogs/day) C Faloutsos, CSD Data Mining Wikipedia change archive (1TB), Fly embryo images (1.5TB), links from Yahoo web S Vogel, LTI Machine Translation Pre-filtered N-gram language model based on statistics on word alignment, 100 TB J Baker, LTI Machine Translation Spoken language recording archive, many languages, many sources, up to 1PB B Becker, RI Computer Vision Social network image/video archive for training computer vision systems, 1-5TB 3
4 CMU PDL History of Scalable Storage 1995 DARPA funds Network-Attached Secure Disks (NASD) NASD spin offs Object Storage Device standardized by T10/SCSI 2004, 2009 Panasas parallel storage system, Gibson co-founder & CTO Primary storage on first petascale computer: LANL Roadrunner Also: NIH, Citadel, ING, BNP, BP, ConocoPhillips, PetroChina, StatOil, Ferrari, BMW, 3M, Lockheed Martin, Northrop Grumman, Sandia, NASA Lustre Linux open source parallel file system With Panasas, Lustre & PVFS, 3/4 top500.org are object-based Graduates go to storage, server & internet companies Eg Google FileSystem (2003) & BigTable (2006) cloud database Parallel NFS achieves IETF RFC in 2010 spurred on by Panasas Linux adoption in , 3.0 and 3.1 (2011) 4
5 For the Experience, Operate A Cloud Two clusters: 3 TF, 2.2 TB, 142 nodes, 1.1K cores, ½ PB Available to CMU escience users as a Hadoop queue IR, ML classes ML research Comp bio research Astro research Geo research Malware analysis Social network analysis Systems research 6x 10GE trunk SFP+ twinax 48 port 10-GE switch 38 10GE links SFP+ twinax Logical rack: 32 worker nodes 6 RAID protected storage nodes PDL & OpenCirrus 6x 10GE trunk SFP+ twinax 48 port 10-GE switch 39 10GE links SFP+ twinax Logical rack: 32 worker nodes 7 RAID protected storage nodes External switch 2 x 10GE SR optical link 24 port 10-GE switch Other OpenCloud sites 10 Gbps LR optical link to NLR 2x 10GE trunk Switch 1-GE down/ 10-GE up 20 x 2 x 1GE links Logical rack: 20 worker nodes 2x 10GE trunk Switch 1-GE down/ 10-GE up 19 x 2 x 1GE links Logical rack: 19 worker nodes 2x 10GE trunk Switch 1-GE down/ 10-GE up 20 x 2 x 1GE links Logical rack: 20 worker nodes 2x 10GE trunk Switch 1-GE down/ 10-GE up 19 x 2 x 1GE links Logical rack: 19 worker nodes CMU OpenCloud CMU OpenCirrus 5
6 To Understand: Cloud FS vs. Parallel FS Hadoop s storage, library, HDFS, is replaceable Replace with PVFS, a user-level Parallel FS, to understand differences Buf: Prefetching HDFS: write once Add deep prefetch Map: Layout Stripe Unit::Node Optimized Launch Rep: Replicate data No HW RAID! To be published in SC11, Nov
7 Replication inside a PVFS file PVFS, like most cluster/parallel file systems, assumes RAID HW HDFS, like GoogleFS, does not like scaling of RAID HW Teach PVFS client to internally replicate (Hybrid approx. HDFS) Code is not production quality error path is too hard for academics J 7
8 Interesting Implementation Issues HDFS performance disk-bound by chunk creation PVFS insufficient parallelism in single stream 8
9 Differences Not Visible in Apps OpenCloud Apps Astrophysics Social network analysis Hadoop helps Job scheduler does load balancing Dataset is directory of files 9
10 Scalable Table Stores Inspired by Google s BigTable Reported to SCALE: >76 PB in one database >10 M operation/sec B-tree with giant nodes Data model is dynamic, lots of columns, strings everywhere Writeback of mutations written as sorted, indexed log files Read-misses search all logs: Log-structured Merge Trees Layered on GFS (HDFS) 10
11 Extending a Prior Benchmark Tool Yahoo! Cloud Serving Benchmark (YCSB) tool steady state load of CRUD (create-read-update-delete) operations Command-line parameters DB to use Target throughput Number of threads Workload parameter file R/W mix Record size Data set YCSB client Workload executor Client threads Stats DB client Cloud DB Extensible: define new workloads Extensible: plug in new clients github.com/brianfrankcooper/ycsb [SoCC10] 11
12 Adv. Features of YCSB++ High Ingest Rate Features Deep batch writing Pre-splitting tablets (given future insert distribution) Bulk-load: MR format map files externally Read Features Read-after-write: what price eventual consistency? Offloading filtering to servers Security ACLs what performance price? Better interpretation of monitoring Integrate knowledge of services, user jobs (Otus) To be published in SoCC (October 2011)
13 Workload parameter file -! R/W mix! -! RecordSize! -! DataSet! -!! Extensions YCSB++ Framework Command-line parameters (e.g, DB name, NumThreads)! YCSB Client (with our extensions) Workload Executor New workloads Multi-Phase Processing Client Threads Stats DB Clients YCSB metrics API ext HBase IcyTable Other DBs Accumulo Client nodes YCSB Client Coordination ZooKeeper-based barrier sync and event notification Ganglia monitoring Hadoop, HDFS and OS metrics Storage Servers github.com/milopolte/ycsb (pushing to main branch) 13
14 Accepted into Apache Incubator Sept
15 Extensions for Monitoring (Otus) Virtual Memory (Bytes) Running Map Tasks Read Requests (ops) CPU Usage HDFS DataNode Read Request From Remote Clients Other Processes Other MRJob Data Node Task Tracker HDFS DataNode CPU Usage Service stats (Hadoop, Hbase, HDFS, ) Walk process group tree looking for specific command lines Aggregate stats for subgroups Customizable displays github.com/otus/otus 15
16 Server side Filtering DoD BigTable HBase Filtering when little data is desired leads to excessive prefetching on server, because it fills scanner batch Size scanner batch to expected result size (scaled buffer) Hbase table was decomposed into more columnar stores, so Accumulo does more work 16
17 Batch Writers & Eventual Consistency Small batches burn excessive client CPU, limiting thruput Large batches saturate servers, limiting benefit of batch 17
18 Batch Writers & Eventual Consistency Deferred write wins, but visible latency can be 100 secs Fraction of requests Fraction of requests (a) HBase: Time lag for different buffer sizes 10 KB 100 KB 1 MB 10 MB 1 (b) IcyTable: 10 Time 100 lag for 1000 different buffer sizes read-after-write time lag (ms) 10 KB 100 KB 1 MB 10 MB read-after-write time lag (ms) 18
19 Pre- (and post-) Tablet Splitting 6 servers Per server: Preload 1M rows; Load 8M rows; Measure@100 ops/s 20% faster load if pre-split post-load rebalancing hurts for minutes 19
20 Improving Ingest Speed: Bulk Load Faster ingest is format with MapReduce, ingest/import with bulk load, rebalance during measurement phase Test: preload, monitor/measure, format bulk, bulk load, monitor/measure, sleep 5 minutes, monitor/measure Per server: Preload 1M rows; Load 8M rows; Measure@100 ops/s Import turns out to be nearly instant, but rebalancing is not Load 48M rows one at a time: secs, mins Bulk load, including formatting time: 5-12 mins (2-5X faster) Data becomes available Queries may slow down Map Reduce Import Rebalancing End-to-end ingest time 20
21 Scaling & Bulk Loading 1/8M rows per server Accumulo (1) (3) (4) (6) 54 Servers (36MF) Servers (36MF) Scaling MR means more files & more compaction Data becomes available Queries may slow down 6 Servers (36MF) Map Reduce Import Rebalancing End-to-end ingest time Minutes PreLoad PL-Rebalance BulkLoad BL-Rebalance 21
22 Rebalancing Timeline (54 Servers/36 MapFiles) (1) (4) (6) (7) (8) (3) Phase 1 rebalancing starts late Too much rebalancing work 22
23 So How Do We Test At Scale? At Cloud scale very few users can afford extended experiment time on public clouds Many systems experiments desire: repeatable, isolated, instrumented, fault-injected, specialized kernels Almost no one running a public cloud could (would) (SHOULD) support such invasive apps 23
24 LANL was going to trash this! 24
25 NSF PRObE to the Rescue NSF Funds the New Mexico Consortium to recycle LANL supercomputers PRObE: Parallel Reconfigurable Observational Environment Low level systems research facility Days to weeks of dedicated usage Complete control of hardware and software Fault injection and failure statistics 25
26 PRObE Hardware Plan Spring 2012: Sitka (2048 cores) acquired 1024 Nodes, Dual Socket, Single Core AMD Opteron; 4GB RAM per core; Full fat-tree Myrinet Summer 2012: Kodiak (2048 cores) acquired 1024 Nodes, Dual Socket, Single Core AMD Opteron, 4GB RAM per core; Fat-tree SDR Infiniband 128 Nodes version at CMU, Marmot, standing up now Fall 2011: Susitna (1700 Cores) being acquired 26 Nodes, 16 core CPUs, 1 GB RAM / core, QDR Infiniband, GPU Planning to build at CMU soon Fall 2013: Nome (1600 cores) anticipated 200 Nodes, Quad Socket, Dual Core AMD Opteron; 2GB RAM per core, Fat-tree DDR Infiniband Fall 2013: Matanuska (3456 Cores) anticipated 36 Nodes, 24 core CPUs, 1-2GB RAM / core, 100Gbit Ethernet 26
27 PRObE Software First, none is allowed Researchers can put any software they want onto the clusters Second, a well known tool managing clusters of hardware for research Emulab ( Flux Group, U. Utah On staging clusters, also on large clusters Enhanced for PRObE hardware, scale, networks, resource partitioning policies, remote power and console, failure injection, deep instrumentation PRObE provides hardware support (spares) 27 Garth Gibson, Oct 2010!
28 For Systems Research Users NSF who can apply rules Includes international and corporate research projects ( best in partnership with US university) newmexicoconsortium.org/probe 28 Garth Gibson, Oct 2010!
29 On Education Front: BigData Masters Extends MSIT Very Large Information Systems (VLIS) Tracks for BigData systems and applications One year on campus, incl. two project courses, plus 7 month internship at end Already using Hadoop on OpenCloud cluster in some courses Systems courses: Distr d Computing, Storage Systems, Cloud Computing, Data Mining, Parallel Comp. Arch & Programming Applications courses: VLIS, Software Eng., Machine Learning, Information Retrieval Seeking students, internship & permanent employers It s all about expanding training of BigData professionals 29
30 Research Sponsors Companies of Parallel Data Consortium: APC, EMC, Facebook, Google, Hewlett-Packard, Hitachi, Intel, Microsoft, NEC, NetApp, Oracle, Panasas, Riverbed, Samsung, Seagate, STEC, Symantec, VMware
YCSB++ Benchmarking Tool Performance Debugging Advanced Features of Scalable Table Stores
YCSB++ Benchmarking Tool Performance Debugging Advanced Features of Scalable Table Stores Swapnil Patil Milo Polte, Wittawat Tantisiriroj, Kai Ren, Lin Xiao, Julio Lopez, Garth Gibson, Adam Fuchs *, Billie
More informationYCSB++ benchmarking tool Performance debugging advanced features of scalable table stores
YCSB++ benchmarking tool Performance debugging advanced features of scalable table stores Swapnil Patil M. Polte, W. Tantisiriroj, K. Ren, L.Xiao, J. Lopez, G.Gibson, A. Fuchs *, B. Rinaldi * Carnegie
More informationCrossing the Chasm: Sneaking a parallel file system into Hadoop
Crossing the Chasm: Sneaking a parallel file system into Hadoop Wittawat Tantisiriroj Swapnil Patil, Garth Gibson PARALLEL DATA LABORATORY Carnegie Mellon University In this work Compare and contrast large
More informationCrossing the Chasm: Sneaking a parallel file system into Hadoop
Crossing the Chasm: Sneaking a parallel file system into Hadoop Wittawat Tantisiriroj Swapnil Patil, Garth Gibson PARALLEL DATA LABORATORY Carnegie Mellon University In this work Compare and contrast large
More informationStorage Systems for Shingled Disks
Storage Systems for Shingled Disks Garth Gibson Carnegie Mellon University and Panasas Inc Anand Suresh, Jainam Shah, Xu Zhang, Swapnil Patil, Greg Ganger Kryder s Law for Magnetic Disks Market expects
More informationCloud Storage and Parallel File Systems
Cloud Storage and Parallel File Systems SNI Storage Developer Conference (SDC09), Sept 2009 Garth Gibson Carnegie Mellon University and Panasas Inc garth@cs.cmu.edu and garth@panasas.com irth of RID (1985-1991)
More informationFailure in Supercomputers in the Post-Terascale Era
Failure in Supercomputers in the Post-Terascale Era Thanks to: Garth Gibson, Carnegie Mellon University and Panasas Inc. DOE SciDAC Petascale Data Storage Institute (PDSI), www.pdsi-scidac.org w/ Bianca
More informationQing Zheng Lin Xiao, Kai Ren, Garth Gibson
Qing Zheng Lin Xiao, Kai Ren, Garth Gibson File System Architecture APP APP APP APP APP APP APP APP metadata operations Metadata Service I/O operations [flat namespace] Low-level Storage Infrastructure
More informationDistributed File Systems II
Distributed File Systems II To do q Very-large scale: Google FS, Hadoop FS, BigTable q Next time: Naming things GFS A radically new environment NFS, etc. Independence Small Scale Variety of workloads Cooperation
More informationDiskReduce: Making Room for More Data on DISCs. Wittawat Tantisiriroj
DiskReduce: Making Room for More Data on DISCs Wittawat Tantisiriroj Lin Xiao, Bin Fan, and Garth Gibson PARALLEL DATA LABORATORY Carnegie Mellon University GFS/HDFS Triplication GFS & HDFS triplicate
More informationStructuring PLFS for Extensibility
Structuring PLFS for Extensibility Chuck Cranor, Milo Polte, Garth Gibson PARALLEL DATA LABORATORY Carnegie Mellon University What is PLFS? Parallel Log Structured File System Interposed filesystem b/w
More informationDistributed Systems 16. Distributed File Systems II
Distributed Systems 16. Distributed File Systems II Paul Krzyzanowski pxk@cs.rutgers.edu 1 Review NFS RPC-based access AFS Long-term caching CODA Read/write replication & disconnected operation DFS AFS
More informationCloud Computing at Yahoo! Thomas Kwan Director, Research Operations Yahoo! Labs
Cloud Computing at Yahoo! Thomas Kwan Director, Research Operations Yahoo! Labs Overview Cloud Strategy Cloud Services Cloud Research Partnerships - 2 - Yahoo! Cloud Strategy 1. Optimizing for Yahoo-scale
More informationSpark Over RDMA: Accelerate Big Data SC Asia 2018 Ido Shamay Mellanox Technologies
Spark Over RDMA: Accelerate Big Data SC Asia 2018 Ido Shamay 1 Apache Spark - Intro Spark within the Big Data ecosystem Data Sources Data Acquisition / ETL Data Storage Data Analysis / ML Serving 3 Apache
More informationExa Scale FSIO Can we get there? Can we afford to?
Exa Scale FSIO Can we get there? Can we afford to? 05/2011 Gary Grider, LANL LA UR 10 04611 Pop Quiz: How Old is this Guy? Back to Exa FSIO Mission Drivers Power is a Driving Issue Power per flop Power
More informationMap Reduce. Yerevan.
Map Reduce Erasmus+ @ Yerevan dacosta@irit.fr Divide and conquer at PaaS 100 % // Typical problem Iterate over a large number of records Extract something of interest from each Shuffle and sort intermediate
More informationData-intensive File Systems for Internet Services: A Rose by Any Other Name... (CMU-PDL )
Carnegie Mellon University Research Showcase @ CMU Parallel Data Laboratory Research Centers and Institutes 10-2008 Data-intensive File Systems for Internet Services: A Rose by Any Other Name... (CMU-PDL-08-114)
More information18-hdfs-gfs.txt Thu Oct 27 10:05: Notes on Parallel File Systems: HDFS & GFS , Fall 2011 Carnegie Mellon University Randal E.
18-hdfs-gfs.txt Thu Oct 27 10:05:07 2011 1 Notes on Parallel File Systems: HDFS & GFS 15-440, Fall 2011 Carnegie Mellon University Randal E. Bryant References: Ghemawat, Gobioff, Leung, "The Google File
More informationHDFS: Hadoop Distributed File System. CIS 612 Sunnie Chung
HDFS: Hadoop Distributed File System CIS 612 Sunnie Chung What is Big Data?? Bulk Amount Unstructured Introduction Lots of Applications which need to handle huge amount of data (in terms of 500+ TB per
More informationNext-Generation Cloud Platform
Next-Generation Cloud Platform Jangwoo Kim Jun 24, 2013 E-mail: jangwoo@postech.ac.kr High Performance Computing Lab Department of Computer Science & Engineering Pohang University of Science and Technology
More informationMapReduce. U of Toronto, 2014
MapReduce U of Toronto, 2014 http://www.google.org/flutrends/ca/ (2012) Average Searches Per Day: 5,134,000,000 2 Motivation Process lots of data Google processed about 24 petabytes of data per day in
More information18-hdfs-gfs.txt Thu Nov 01 09:53: Notes on Parallel File Systems: HDFS & GFS , Fall 2012 Carnegie Mellon University Randal E.
18-hdfs-gfs.txt Thu Nov 01 09:53:32 2012 1 Notes on Parallel File Systems: HDFS & GFS 15-440, Fall 2012 Carnegie Mellon University Randal E. Bryant References: Ghemawat, Gobioff, Leung, "The Google File
More informationCA485 Ray Walshe Google File System
Google File System Overview Google File System is scalable, distributed file system on inexpensive commodity hardware that provides: Fault Tolerance File system runs on hundreds or thousands of storage
More informationFeedback on BeeGFS. A Parallel File System for High Performance Computing
Feedback on BeeGFS A Parallel File System for High Performance Computing Philippe Dos Santos et Georges Raseev FR 2764 Fédération de Recherche LUmière MATière December 13 2016 LOGO CNRS LOGO IO December
More informationIntroduction to Hadoop. Owen O Malley Yahoo!, Grid Team
Introduction to Hadoop Owen O Malley Yahoo!, Grid Team owen@yahoo-inc.com Who Am I? Yahoo! Architect on Hadoop Map/Reduce Design, review, and implement features in Hadoop Working on Hadoop full time since
More informationECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective
ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective Part II: Data Center Software Architecture: Topic 1: Distributed File Systems GFS (The Google File System) 1 Filesystems
More informationMapR Enterprise Hadoop
2014 MapR Technologies 2014 MapR Technologies 1 MapR Enterprise Hadoop Top Ranked Cloud Leaders 500+ Customers 2014 MapR Technologies 2 Key MapR Advantage Partners Business Services APPLICATIONS & OS ANALYTICS
More informationBigData and Map Reduce VITMAC03
BigData and Map Reduce VITMAC03 1 Motivation Process lots of data Google processed about 24 petabytes of data per day in 2009. A single machine cannot serve all the data You need a distributed system to
More informationFusion iomemory PCIe Solutions from SanDisk and Sqrll make Accumulo Hypersonic
WHITE PAPER Fusion iomemory PCIe Solutions from SanDisk and Sqrll make Accumulo Hypersonic Western Digital Technologies, Inc. 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Executive
More informationDistributed Filesystem
Distributed Filesystem 1 How do we get data to the workers? NAS Compute Nodes SAN 2 Distributing Code! Don t move data to workers move workers to the data! - Store data on the local disks of nodes in the
More informationBig Data Analytics. Izabela Moise, Evangelos Pournaras, Dirk Helbing
Big Data Analytics Izabela Moise, Evangelos Pournaras, Dirk Helbing Izabela Moise, Evangelos Pournaras, Dirk Helbing 1 Big Data "The world is crazy. But at least it s getting regular analysis." Izabela
More informationTopics. Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples
Hadoop Introduction 1 Topics Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples 2 Big Data Analytics What is Big Data?
More informationCluster Setup and Distributed File System
Cluster Setup and Distributed File System R&D Storage for the R&D Storage Group People Involved Gaetano Capasso - INFN-Naples Domenico Del Prete INFN-Naples Diacono Domenico INFN-Bari Donvito Giacinto
More informationLecture 11 Hadoop & Spark
Lecture 11 Hadoop & Spark Dr. Wilson Rivera ICOM 6025: High Performance Computing Electrical and Computer Engineering Department University of Puerto Rico Outline Distributed File Systems Hadoop Ecosystem
More informationIntroduction to MapReduce
Basics of Cloud Computing Lecture 4 Introduction to MapReduce Satish Srirama Some material adapted from slides by Jimmy Lin, Christophe Bisciglia, Aaron Kimball, & Sierra Michels-Slettvet, Google Distributed
More informationMixApart: Decoupled Analytics for Shared Storage Systems. Madalin Mihailescu, Gokul Soundararajan, Cristiana Amza University of Toronto and NetApp
MixApart: Decoupled Analytics for Shared Storage Systems Madalin Mihailescu, Gokul Soundararajan, Cristiana Amza University of Toronto and NetApp Hadoop Pig, Hive Hadoop + Enterprise storage?! Shared storage
More informationA BigData Tour HDFS, Ceph and MapReduce
A BigData Tour HDFS, Ceph and MapReduce These slides are possible thanks to these sources Jonathan Drusi - SCInet Toronto Hadoop Tutorial, Amir Payberah - Course in Data Intensive Computing SICS; Yahoo!
More informationData Intensive Scalable Computing. Thanks to: Randal E. Bryant Carnegie Mellon University
Data Intensive Scalable Computing Thanks to: Randal E. Bryant Carnegie Mellon University http://www.cs.cmu.edu/~bryant Big Data Sources: Seismic Simulations Wave propagation during an earthquake Large-scale
More informationThe Hadoop Distributed File System Konstantin Shvachko Hairong Kuang Sanjay Radia Robert Chansler
The Hadoop Distributed File System Konstantin Shvachko Hairong Kuang Sanjay Radia Robert Chansler MSST 10 Hadoop in Perspective Hadoop scales computation capacity, storage capacity, and I/O bandwidth by
More informationScalable Web Programming. CS193S - Jan Jannink - 2/25/10
Scalable Web Programming CS193S - Jan Jannink - 2/25/10 Weekly Syllabus 1.Scalability: (Jan.) 2.Agile Practices 3.Ecology/Mashups 4.Browser/Client 7.Analytics 8.Cloud/Map-Reduce 9.Published APIs: (Mar.)*
More informationHadoop 2.x Core: YARN, Tez, and Spark. Hortonworks Inc All Rights Reserved
Hadoop 2.x Core: YARN, Tez, and Spark YARN Hadoop Machine Types top-of-rack switches core switch client machines have client-side software used to access a cluster to process data master nodes run Hadoop
More informationDiskReduce: Making Room for More Data on DISCs. Wittawat Tantisiriroj
DiskReduce: Making Room for More Data on DISCs Wittawat Tantisiriroj Lin Xiao, in Fan, and Garth Gibson PARALLEL DATA LAORATORY Carnegie Mellon University GFS/HDFS Triplication GFS & HDFS triplicate every
More informationCloud Computing and Hadoop Distributed File System. UCSB CS170, Spring 2018
Cloud Computing and Hadoop Distributed File System UCSB CS70, Spring 08 Cluster Computing Motivations Large-scale data processing on clusters Scan 000 TB on node @ 00 MB/s = days Scan on 000-node cluster
More informationThe Fusion Distributed File System
Slide 1 / 44 The Fusion Distributed File System Dongfang Zhao February 2015 Slide 2 / 44 Outline Introduction FusionFS System Architecture Metadata Management Data Movement Implementation Details Unique
More informationCS 61C: Great Ideas in Computer Architecture. MapReduce
CS 61C: Great Ideas in Computer Architecture MapReduce Guest Lecturer: Justin Hsia 3/06/2013 Spring 2013 Lecture #18 1 Review of Last Lecture Performance latency and throughput Warehouse Scale Computing
More informationNFS, GPFS, PVFS, Lustre Batch-scheduled systems: Clusters, Grids, and Supercomputers Programming paradigm: HPC, MTC, and HTC
Segregated storage and compute NFS, GPFS, PVFS, Lustre Batch-scheduled systems: Clusters, Grids, and Supercomputers Programming paradigm: HPC, MTC, and HTC Co-located storage and compute HDFS, GFS Data
More informationVoldemort. Smruti R. Sarangi. Department of Computer Science Indian Institute of Technology New Delhi, India. Overview Design Evaluation
Voldemort Smruti R. Sarangi Department of Computer Science Indian Institute of Technology New Delhi, India Smruti R. Sarangi Leader Election 1/29 Outline 1 2 3 Smruti R. Sarangi Leader Election 2/29 Data
More informationOracle NoSQL Database and Cisco- Collaboration that produces results. 1 Copyright 2011, Oracle and/or its affiliates. All rights reserved.
Oracle NoSQL Database and Cisco- Collaboration that produces results 1 Copyright 2011, Oracle and/or its affiliates. All rights reserved. What is Big Data? SOCIAL BLOG SMART METER VOLUME VELOCITY VARIETY
More informationHadoop File System S L I D E S M O D I F I E D F R O M P R E S E N T A T I O N B Y B. R A M A M U R T H Y 11/15/2017
Hadoop File System 1 S L I D E S M O D I F I E D F R O M P R E S E N T A T I O N B Y B. R A M A M U R T H Y Moving Computation is Cheaper than Moving Data Motivation: Big Data! What is BigData? - Google
More informationMAHA. - Supercomputing System for Bioinformatics
MAHA - Supercomputing System for Bioinformatics - 2013.01.29 Outline 1. MAHA HW 2. MAHA SW 3. MAHA Storage System 2 ETRI HPC R&D Area - Overview Research area Computing HW MAHA System HW - Rpeak : 0.3
More informationCISC 7610 Lecture 2b The beginnings of NoSQL
CISC 7610 Lecture 2b The beginnings of NoSQL Topics: Big Data Google s infrastructure Hadoop: open google infrastructure Scaling through sharding CAP theorem Amazon s Dynamo 5 V s of big data Everyone
More informationCIS 601 Graduate Seminar. Dr. Sunnie S. Chung Dhruv Patel ( ) Kalpesh Sharma ( )
Guide: CIS 601 Graduate Seminar Presented By: Dr. Sunnie S. Chung Dhruv Patel (2652790) Kalpesh Sharma (2660576) Introduction Background Parallel Data Warehouse (PDW) Hive MongoDB Client-side Shared SQL
More informationBIG DATA AND HADOOP ON THE ZFS STORAGE APPLIANCE
BIG DATA AND HADOOP ON THE ZFS STORAGE APPLIANCE BRETT WENINGER, MANAGING DIRECTOR 10/21/2014 ADURANT APPROACH TO BIG DATA Align to Un/Semi-structured Data Instead of Big Scale out will become Big Greatest
More informationCan Parallel Replication Benefit Hadoop Distributed File System for High Performance Interconnects?
Can Parallel Replication Benefit Hadoop Distributed File System for High Performance Interconnects? N. S. Islam, X. Lu, M. W. Rahman, and D. K. Panda Network- Based Compu2ng Laboratory Department of Computer
More informationDatabase Architecture 2 & Storage. Instructor: Matei Zaharia cs245.stanford.edu
Database Architecture 2 & Storage Instructor: Matei Zaharia cs245.stanford.edu Summary from Last Time System R mostly matched the architecture of a modern RDBMS» SQL» Many storage & access methods» Cost-based
More informationNew Oracle NoSQL Database APIs that Speed Insertion and Retrieval
New Oracle NoSQL Database APIs that Speed Insertion and Retrieval O R A C L E W H I T E P A P E R F E B R U A R Y 2 0 1 6 1 NEW ORACLE NoSQL DATABASE APIs that SPEED INSERTION AND RETRIEVAL Introduction
More informationIndexFS: Scaling File System Metadata Performance with Stateless Caching and Bulk Insertion
IndexFS: Scaling File System Metadata Performance with Stateless Caching and Bulk Insertion Kai Ren Qing Zheng, Swapnil Patil, Garth Gibson PARALLEL DATA LABORATORY Carnegie Mellon University Why Scalable
More informationSun Lustre Storage System Simplifying and Accelerating Lustre Deployments
Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments Torben Kling-Petersen, PhD Presenter s Name Principle Field Title andengineer Division HPC &Cloud LoB SunComputing Microsystems
More informationChapter 5. The MapReduce Programming Model and Implementation
Chapter 5. The MapReduce Programming Model and Implementation - Traditional computing: data-to-computing (send data to computing) * Data stored in separate repository * Data brought into system for computing
More informationArchitecture of a Real-Time Operational DBMS
Architecture of a Real-Time Operational DBMS Srini V. Srinivasan Founder, Chief Development Officer Aerospike CMG India Keynote Thane December 3, 2016 [ CMGI Keynote, Thane, India. 2016 Aerospike Inc.
More informationYuval Carmel Tel-Aviv University "Advanced Topics in Storage Systems" - Spring 2013
Yuval Carmel Tel-Aviv University "Advanced Topics in About & Keywords Motivation & Purpose Assumptions Architecture overview & Comparison Measurements How does it fit in? The Future 2 About & Keywords
More informationHow Apache Hadoop Complements Existing BI Systems. Dr. Amr Awadallah Founder, CTO Cloudera,
How Apache Hadoop Complements Existing BI Systems Dr. Amr Awadallah Founder, CTO Cloudera, Inc. Twitter: @awadallah, @cloudera 2 The Problems with Current Data Systems BI Reports + Interactive Apps RDBMS
More informationCS November 2017
Bigtable Highly available distributed storage Distributed Systems 18. Bigtable Built with semi-structured data in mind URLs: content, metadata, links, anchors, page rank User data: preferences, account
More informationCLOUD-SCALE FILE SYSTEMS
Data Management in the Cloud CLOUD-SCALE FILE SYSTEMS 92 Google File System (GFS) Designing a file system for the Cloud design assumptions design choices Architecture GFS Master GFS Chunkservers GFS Clients
More informationMap-Reduce. Marco Mura 2010 March, 31th
Map-Reduce Marco Mura (mura@di.unipi.it) 2010 March, 31th This paper is a note from the 2009-2010 course Strumenti di programmazione per sistemi paralleli e distribuiti and it s based by the lessons of
More informationMapReduce, Hadoop and Spark. Bompotas Agorakis
MapReduce, Hadoop and Spark Bompotas Agorakis Big Data Processing Most of the computations are conceptually straightforward on a single machine but the volume of data is HUGE Need to use many (1.000s)
More informationWelcome to the New Era of Cloud Computing
Welcome to the New Era of Cloud Computing Aaron Kimball The web is replacing the desktop 1 SDKs & toolkits are there What about the backend? Image: Wikipedia user Calyponte 2 Two key concepts Processing
More information2013 AWS Worldwide Public Sector Summit Washington, D.C.
2013 AWS Worldwide Public Sector Summit Washington, D.C. EMR for Fun and for Profit Ben Butler Sr. Manager, Big Data butlerb@amazon.com @bensbutler Overview 1. What is big data? 2. What is AWS Elastic
More informationGoogle File System (GFS) and Hadoop Distributed File System (HDFS)
Google File System (GFS) and Hadoop Distributed File System (HDFS) 1 Hadoop: Architectural Design Principles Linear scalability More nodes can do more work within the same time Linear on data size, linear
More informationCPSC 426/526. Cloud Computing. Ennan Zhai. Computer Science Department Yale University
CPSC 426/526 Cloud Computing Ennan Zhai Computer Science Department Yale University Recall: Lec-7 In the lec-7, I talked about: - P2P vs Enterprise control - Firewall - NATs - Software defined network
More informationDistributed Systems CS6421
Distributed Systems CS6421 Intro to Distributed Systems and the Cloud Prof. Tim Wood v I teach: Software Engineering, Operating Systems, Sr. Design I like: distributed systems, networks, building cool
More informationLet s Make Parallel File System More Parallel
Let s Make Parallel File System More Parallel [LA-UR-15-25811] Qing Zheng 1, Kai Ren 1, Garth Gibson 1, Bradley W. Settlemyer 2 1 Carnegie MellonUniversity 2 Los AlamosNationalLaboratory HPC defined by
More information10 Million Smart Meter Data with Apache HBase
10 Million Smart Meter Data with Apache HBase 5/31/2017 OSS Solution Center Hitachi, Ltd. Masahiro Ito OSS Summit Japan 2017 Who am I? Masahiro Ito ( 伊藤雅博 ) Software Engineer at Hitachi, Ltd. Focus on
More informationThe Leading Parallel Cluster File System
The Leading Parallel Cluster File System www.thinkparq.com www.beegfs.io ABOUT BEEGFS What is BeeGFS BeeGFS (formerly FhGFS) is the leading parallel cluster file system, developed with a strong focus on
More informationKonstantin Shvachko, Hairong Kuang, Sanjay Radia, Robert Chansler Yahoo! Sunnyvale, California USA {Shv, Hairong, SRadia,
Konstantin Shvachko, Hairong Kuang, Sanjay Radia, Robert Chansler Yahoo! Sunnyvale, California USA {Shv, Hairong, SRadia, Chansler}@Yahoo-Inc.com Presenter: Alex Hu } Introduction } Architecture } File
More informationCorrelX: A Cloud-Based VLBI Correlator
CorrelX: A Cloud-Based VLBI Correlator V. Pankratius, A. J. Vazquez, P. Elosegui Massachusetts Institute of Technology Haystack Observatory pankrat@mit.edu, victorpankratius.com 5 th International VLBI
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung December 2003 ACM symposium on Operating systems principles Publisher: ACM Nov. 26, 2008 OUTLINE INTRODUCTION DESIGN OVERVIEW
More informationThe BioHPC Nucleus Cluster & Future Developments
1 The BioHPC Nucleus Cluster & Future Developments Overview Today we ll talk about the BioHPC Nucleus HPC cluster with some technical details for those interested! How is it designed? What hardware does
More informationCS November 2018
Bigtable Highly available distributed storage Distributed Systems 19. Bigtable Built with semi-structured data in mind URLs: content, metadata, links, anchors, page rank User data: preferences, account
More informationOracle Exadata X7. Uwe Kirchhoff Oracle ACS - Delivery Senior Principal Service Delivery Engineer
Oracle Exadata X7 Uwe Kirchhoff Oracle ACS - Delivery Senior Principal Service Delivery Engineer 05.12.2017 Oracle Engineered Systems ZFS Backup Appliance Zero Data Loss Recovery Appliance Exadata Database
More informationThe Google File System. Alexandru Costan
1 The Google File System Alexandru Costan Actions on Big Data 2 Storage Analysis Acquisition Handling the data stream Data structured unstructured semi-structured Results Transactions Outline File systems
More informationName: Instructions. Problem 1 : Short answer. [48 points] CMU / Storage Systems 23 Feb 2011 Spring 2012 Exam 1
CMU 18-746/15-746 Storage Systems 23 Feb 2011 Spring 2012 Exam 1 Instructions Name: There are three (3) questions on the exam. You may find questions that could have several answers and require an explanation
More informationHadoop An Overview. - Socrates CCDH
Hadoop An Overview - Socrates CCDH What is Big Data? Volume Not Gigabyte. Terabyte, Petabyte, Exabyte, Zettabyte - Due to handheld gadgets,and HD format images and videos - In total data, 90% of them collected
More informationLazyBase: Trading freshness and performance in a scalable database
LazyBase: Trading freshness and performance in a scalable database (EuroSys 2012) Jim Cipar, Greg Ganger, *Kimberly Keeton, *Craig A. N. Soules, *Brad Morrey, *Alistair Veitch PARALLEL DATA LABORATORY
More informationParallel File Systems for HPC
Introduction to Scuola Internazionale Superiore di Studi Avanzati Trieste November 2008 Advanced School in High Performance and Grid Computing Outline 1 The Need for 2 The File System 3 Cluster & A typical
More informationGoogle File System and BigTable. and tiny bits of HDFS (Hadoop File System) and Chubby. Not in textbook; additional information
Subject 10 Fall 2015 Google File System and BigTable and tiny bits of HDFS (Hadoop File System) and Chubby Not in textbook; additional information Disclaimer: These abbreviated notes DO NOT substitute
More informationVOLTDB + HP VERTICA. page
VOLTDB + HP VERTICA ARCHITECTURE FOR FAST AND BIG DATA ARCHITECTURE FOR FAST + BIG DATA FAST DATA Fast Serve Analytics BIG DATA BI Reporting Fast Operational Database Streaming Analytics Columnar Analytics
More informationScalable I/O, File Systems, and Storage Networks R&D at Los Alamos LA-UR /2005. Gary Grider CCN-9
Scalable I/O, File Systems, and Storage Networks R&D at Los Alamos LA-UR-05-2030 05/2005 Gary Grider CCN-9 Background Disk2500 TeraBytes Parallel I/O What drives us? Provide reliable, easy-to-use, high-performance,
More informationCS6030 Cloud Computing. Acknowledgements. Today s Topics. Intro to Cloud Computing 10/20/15. Ajay Gupta, WMU-CS. WiSe Lab
CS6030 Cloud Computing Ajay Gupta B239, CEAS Computer Science Department Western Michigan University ajay.gupta@wmich.edu 276-3104 1 Acknowledgements I have liberally borrowed these slides and material
More informationNext-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads
Next-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads Liran Zvibel CEO, Co-founder WekaIO @liranzvibel 1 WekaIO Matrix: Full-featured and Flexible Public or Private S3 Compatible
More informationStorage Optimization with Oracle Database 11g
Storage Optimization with Oracle Database 11g Terabytes of Data Reduce Storage Costs by Factor of 10x Data Growth Continues to Outpace Budget Growth Rate of Database Growth 1000 800 600 400 200 1998 2000
More informationEsgynDB Enterprise 2.0 Platform Reference Architecture
EsgynDB Enterprise 2.0 Platform Reference Architecture This document outlines a Platform Reference Architecture for EsgynDB Enterprise, built on Apache Trafodion (Incubating) implementation with licensed
More informationCIT 668: System Architecture. Amazon Web Services
CIT 668: System Architecture Amazon Web Services Topics 1. AWS Global Infrastructure 2. Foundation Services 1. Compute 2. Storage 3. Database 4. Network 3. AWS Economics Amazon Services Architecture Regions
More informationReflections on Failure in Post-Terascale Parallel Computing
Reflections on Failure in Post-Terascale Parallel Computing 2007 Int. Conf. on Parallel Processing, Xi An China Garth Gibson Carnegie Mellon University and Panasas Inc. DOE SciDAC Petascale Data Storage
More informationWhere We Are. Review: Parallel DBMS. Parallel DBMS. Introduction to Data Management CSE 344
Where We Are Introduction to Data Management CSE 344 Lecture 22: MapReduce We are talking about parallel query processing There exist two main types of engines: Parallel DBMSs (last lecture + quick review)
More informationIntroduction to Data Management CSE 344
Introduction to Data Management CSE 344 Lecture 24: MapReduce CSE 344 - Winter 215 1 HW8 MapReduce (Hadoop) w/ declarative language (Pig) Due next Thursday evening Will send out reimbursement codes later
More informationIn the multi-core age, How do larger, faster and cheaper and more responsive memory sub-systems affect data management? Dhabaleswar K.
In the multi-core age, How do larger, faster and cheaper and more responsive sub-systems affect data management? Panel at ADMS 211 Dhabaleswar K. (DK) Panda Network-Based Computing Laboratory Department
More informationAdvanced Database Systems
Lecture II Storage Layer Kyumars Sheykh Esmaili Course s Syllabus Core Topics Storage Layer Query Processing and Optimization Transaction Management and Recovery Advanced Topics Cloud Computing and Web
More informationParallel Programming Principle and Practice. Lecture 10 Big Data Processing with MapReduce
Parallel Programming Principle and Practice Lecture 10 Big Data Processing with MapReduce Outline MapReduce Programming Model MapReduce Examples Hadoop 2 Incredible Things That Happen Every Minute On The
More informationClustering Lecture 8: MapReduce
Clustering Lecture 8: MapReduce Jing Gao SUNY Buffalo 1 Divide and Conquer Work Partition w 1 w 2 w 3 worker worker worker r 1 r 2 r 3 Result Combine 4 Distributed Grep Very big data Split data Split data
More information