Giza: Erasure Coding Objects across Global Data Centers
|
|
- Angel Joseph
- 6 years ago
- Views:
Transcription
1 Giza: Erasure Coding Objects across Global Data Centers Yu Lin Chen*, Shuai Mu, Jinyang Li, Cheng Huang *, Jin li *, Aaron Ogus *, and Douglas Phillips* New York University, *Microsoft Corporation USENIX ATC, Santa Clara, CA, July
2 Azure data centers span the globe 2
3 Data center fault tolerance: Geo Replication blob 3
4 Data center fault tolerance: Erasure coding blob a b p Project Giza 4
5 The cost of erasure coding a b blob Project Giza 5
6 When is cross DC coding a win? 1. Inexpensive cross DC communication. 2. Coding friendly workload a) Stores a lot of data in large objects (less coding overhead) b) Infrequent reconstruction due to read or failure. 6
7 Cross DC communication is becoming less expensive Largest bandwidth capacity today with 160Tpbs 7
8 In search of a coding friendly workload 3 months trace: January March
9 OneDrive is occupied mostly by large objects 9
10 OneDrive is occupied mostly by large objects Object > 1GB occupies 53% of total storage capacity. 10
11 OneDrive is occupied mostly by large objects Objects < 4MB occupies only 0.9% of total storage capacity 11
12 OneDrive reads are infrequent after a month 12
13 OneDrive reads are infrequent after a month 47% of the total bytes read occur within the first day of the object creation 13
14 OneDrive reads are infrequent after a month less than 2% of the total bytes read occur after a month 14
15 Outline 1. The case for erasure coding across data centers. 2. Giza design and challenges 3. Experimental Results 15
16 Giza: a strongly consistent versioned object store Erasure code each object individually. Optimize latency for Put and Get. Leverage existing single-dc cloud storage API. 16
17 Giza s architecture DC 2 Table Blob DC 1 DC 3 17
18 Giza s workflow Client Put(Blob1, Data) DC 2 Table Blob DC 1 DC 3 18
19 Giza s workflow Client Put(Blob1, Data) Put(GUID1, Fragment A) DC 2 Table Blob DC 1 DC 3 19
20 Giza s workflow Client Put(Blob1, Data) {DC1: GUID1, DC2: GUID2, DC3: GUID3} DC 2 Table Blob DC 1 DC 3 20
21 Giza metadata versioning Let md this = {DC 1 :GUID 4, DC 2 :GUID 5, DC 3 :GUID 6 } Key Metadata (version #: actual metadata) Blob1 v1: md 1 Key Blob1 Metadata (version #: actual metadata) v1: md 1, v2: md this 21
22 Inconsistency during concurrent put operations Client Put(Blob1, Data1) Client Put(Blob1, Data2) DC 2 Table Blob DC 1 DC 3 22
23 Inconsistency during concurrent put operations Order of arrival md2 md1 md2 md1 md2 md1 Key Metadata Key Metadata Key Metadata Blob1 v1: md 1 DC 1 Blob1 v1: md 2 DC 2 Blob1 v1: md 1 DC 3 WARNING: INCONSISTENT STATE 23
24 Inconsistency during concurrent put operations Order of arrival md2 md1 md2 md1 md2 md1 Key Metadata Blob1 v1: md 1 DC 1 Key Metadata Blob1 v1: md 2 DC 2 Key Metadata FAILURE! Blob1 v1: md 1 DC 3 WARNING: INCONSISTENT STATE 24
25 Choosing version value via consensus Key Metadata Key Metadata Key Metadata Blob1 v1: md 1 v2: md 2 Blob1 v1: md 1 v2: md 2 Blob1 v1: md 1 v2: md 2 DC 1 DC 2 OR DC 3 Key Metadata Key Metadata Key Metadata Blob1 v1: md 2 v2: md 1 Blob1 v1: md 2 v2: md 1 Blob1 v1: md 2 v2: md 1 DC 1 DC 2 DC 3 25
26 Paxos and Fast Paxos properties 26
27 Paxos and Fast Paxos properties 27
28 Implementing Fast Paxos with Azure table API Let md this = ({DC 1 :GUID 1, DC 2 :GUID 2, DC 3 :GUID 3 }, Paxos states) DC 2 Table Blob DC 1 DC 3 28
29 Implementing Fast Paxos with Azure table API Let md this = ({DC 1 :GUID 1, DC 2 :GUID 2, DC 3 :GUID 3 }, Paxos states) DC 2 Table Blob DC 1 DC 3 29
30 Implementing Fast Paxos with Azure table API Let md this = ({DC 1 :GUID 1, DC 2 :GUID 2, DC 3 :GUID 3 }, Paxos states) Read metadata row and decide to accept or reject proposal (v2 = md this ) Table Blob Etag = 1 Remote DC 30
31 Implementing Fast Paxos with Azure table API Let md this = ({DC 1 :GUID 1, DC 2 :GUID 2, DC 3 :GUID 3 }, Paxos states) Etag = 1 Fail to update table since etag has changed Write metadata row back to the table if proposal accepted Table Blob Etag = 2 Remote DC 31
32 Implementing Fast Paxos with Azure table API Let md this = ({DC 1 :GUID 1, DC 2 :GUID 2, DC 3 :GUID 3 }, Paxos states) Etag = 1 Successfully update table since etag hasn t changed Write metadata row back to the table if proposal accepted Table Blob Etag = 1 Remote DC 32
33 Parallelizing metadata and data path DC 2 Table Blob DC 1 DC 3 33
34 Data path fails, metadata path succeeds Key Metadata (version #: actual metadata) Blob1 v1: md 1, v2: md 2 Ongoing, data path failed During read, Giza will revert back to the latest version where the data is available. 34
35 Giza Put Execute metadata and data path in parallel. Upon success, return acknowledgement to the client. In the background, update highest committed version number in the metadata row. Key Committed Metadata (version #: actual metadata) Blob1 1 v1: md 1, v2: md 2 Ongoing, not committed 35
36 Giza Get Optimistically read the latest committed version locally. Execute both data path and metadata path in parallel: Data path retrieves the data fragments indicated by the local latest committed version. Metadata path will retrieve the metadata row from all data centers and validate the latest committed version. 36
37 Outline 1. The case for erasure coding across data centers. 2. Giza design and challenges 3. Experimental Results 37
38 Giza deployment configurations US
39 Giza deployment configurations US
40 Giza deployment configurations World
41 Giza deployment configurations World
42 Metadata Path: Fast vs Classic Paxos 23ms 32ms 64ms 42
43 Metadata Path: Fast vs Classic Paxos 102ms 240ms 139ms 43
44 Giza Put Latency (World-2-1) 44
45 Giza Get Latency (World-2-1) 45
46 Summary Recent trend in data centers supports the economic feasibility of erasure coding across data centers. Giza is a strongly consistent versioned object store, built on top of Azure storage. Giza optimizes for latency and is fast in the common case. 46
Erasure Coding in Object Stores: Challenges and Opportunities
Erasure Coding in Object Stores: Challenges and Opportunities Lewis Tseng Boston College July 2018, PODC Acknowledgements Nancy Lynch Muriel Medard Kishori Konwar Prakash Narayana Moorthy Viveck R. Cadambe
More informationJanus. Consolidating Concurrency Control and Consensus for Commits under Conflicts. Shuai Mu, Lamont Nelson, Wyatt Lloyd, Jinyang Li
Janus Consolidating Concurrency Control and Consensus for Commits under Conflicts Shuai Mu, Lamont Nelson, Wyatt Lloyd, Jinyang Li New York University, University of Southern California State of the Art
More informationPelican: A building block for exascale cold data storage
Pelican: A building block for exascale cold data storage Shobana Balarishnan, Richard Black, Austin Donnelly, Paul England, Adam Glass, Dave Harper, Sergey Legtchenko, Aaron Ogus, Eric Peterson, Antony
More informationPricing Intra-Datacenter Networks with
Pricing Intra-Datacenter Networks with Over-Committed Bandwidth Guarantee Jian Guo 1, Fangming Liu 1, Tao Wang 1, and John C.S. Lui 2 1 Cloud Datacenter & Green Computing/Communications Research Group
More informationTape in the Microsoft Datacenter: The Good and Bad of Tape as a Target for Cloud-based Archival Storage
Tape in the Microsoft Datacenter: The Good and Bad of Tape as a Target for Cloud-based Archival Storage Marvin McNett Principal Development Manager Microsoft Azure Storage Content I. Azure Archival Storage
More informationBuilding Consistent Transactions with Inconsistent Replication
Building Consistent Transactions with Inconsistent Replication Irene Zhang, Naveen Kr. Sharma, Adriana Szekeres, Arvind Krishnamurthy, Dan R. K. Ports University of Washington Distributed storage systems
More informationYves Goeleven. Solution Architect - Particular Software. Shipping software since Azure MVP since Co-founder & board member AZUG
Storage Services Yves Goeleven Solution Architect - Particular Software Shipping software since 2001 Azure MVP since 2010 Co-founder & board member AZUG NServiceBus & MessageHandler Used azure storage?
More informationEnhancing Throughput of
Enhancing Throughput of NCA 2017 Zhongmiao Li, Peter Van Roy and Paolo Romano Enhancing Throughput of Partially Replicated State Machines via NCA 2017 Zhongmiao Li, Peter Van Roy and Paolo Romano Enhancing
More informationABSTRACT I. INTRODUCTION
International Journal of Scientific Research in Computer Science, Engineering and Information Technology 2018 IJSRCSEIT Volume 3 Issue 3 ISS: 2456-3307 Hadoop Periodic Jobs Using Data Blocks to Achieve
More informationEverything You Need to Know About MySQL Group Replication
Everything You Need to Know About MySQL Group Replication Luís Soares (luis.soares@oracle.com) Principal Software Engineer, MySQL Replication Lead Copyright 2017, Oracle and/or its affiliates. All rights
More informationECS High Availability Design
ECS High Availability Design March 2018 A Dell EMC white paper Revisions Date Mar 2018 Aug 2017 July 2017 Description Version 1.2 - Updated to include ECS version 3.2 content Version 1.1 - Updated to include
More informationData Centers and Cloud Computing
Data Centers and Cloud Computing CS677 Guest Lecture Tim Wood 1 Data Centers Large server and storage farms 1000s of servers Many TBs or PBs of data Used by Enterprises for server applications Internet
More informationData Centers and Cloud Computing. Slides courtesy of Tim Wood
Data Centers and Cloud Computing Slides courtesy of Tim Wood 1 Data Centers Large server and storage farms 1000s of servers Many TBs or PBs of data Used by Enterprises for server applications Internet
More informationData Centers and Cloud Computing. Data Centers
Data Centers and Cloud Computing Slides courtesy of Tim Wood 1 Data Centers Large server and storage farms 1000s of servers Many TBs or PBs of data Used by Enterprises for server applications Internet
More informationPatterns on XRegional Data Consistency
Patterns on XRegional Data Consistency Contents The problem... 3 Introducing XRegional... 3 The solution... 5 Enabling consistency... 6 The XRegional Framework: A closer look... 8 Some considerations...
More informationREFERENCE ARCHITECTURE Quantum StorNext and Cloudian HyperStore
REFERENCE ARCHITECTURE Quantum StorNext and Cloudian HyperStore CLOUDIAN + QUANTUM REFERENCE ARCHITECTURE 1 Table of Contents Introduction to Quantum StorNext 3 Introduction to Cloudian HyperStore 3 Audience
More informationCS /15/16. Paul Krzyzanowski 1. Question 1. Distributed Systems 2016 Exam 2 Review. Question 3. Question 2. Question 5.
Question 1 What makes a message unstable? How does an unstable message become stable? Distributed Systems 2016 Exam 2 Review Paul Krzyzanowski Rutgers University Fall 2016 In virtual sychrony, a message
More informationNetChain: Scale-Free Sub-RTT Coordination
NetChain: Scale-Free Sub-RTT Coordination Xin Jin Xiaozhou Li, Haoyu Zhang, Robert Soulé, Jeongkeun Lee, Nate Foster, Changhoon Kim, Ion Stoica Conventional wisdom: avoid coordination NetChain: lightning
More informationSDPaxos: Building Efficient Semi-Decentralized Geo-replicated State Machines
SDPaxos: Building Efficient Semi-Decentralized Geo-replicated State Machines Hanyu Zhao *, Quanlu Zhang, Zhi Yang *, Ming Wu, Yafei Dai * * Peking University Microsoft Research Replication for Fault Tolerance
More informationPaxos Replicated State Machines as the Basis of a High- Performance Data Store
Paxos Replicated State Machines as the Basis of a High- Performance Data Store William J. Bolosky, Dexter Bradshaw, Randolph B. Haagens, Norbert P. Kusters and Peng Li March 30, 2011 Q: How to build a
More informationCLOUD-SCALE FILE SYSTEMS
Data Management in the Cloud CLOUD-SCALE FILE SYSTEMS 92 Google File System (GFS) Designing a file system for the Cloud design assumptions design choices Architecture GFS Master GFS Chunkservers GFS Clients
More informationThe Google File System (GFS)
1 The Google File System (GFS) CS60002: Distributed Systems Antonio Bruto da Costa Ph.D. Student, Formal Methods Lab, Dept. of Computer Sc. & Engg., Indian Institute of Technology Kharagpur 2 Design constraints
More informationCMU SCS CMU SCS Who: What: When: Where: Why: CMU SCS
Carnegie Mellon Univ. Dept. of Computer Science 15-415/615 - DB s C. Faloutsos A. Pavlo Lecture#23: Distributed Database Systems (R&G ch. 22) Administrivia Final Exam Who: You What: R&G Chapters 15-22
More informationGFS Overview. Design goals/priorities Design for big-data workloads Huge files, mostly appends, concurrency, huge bandwidth Design for failures
GFS Overview Design goals/priorities Design for big-data workloads Huge files, mostly appends, concurrency, huge bandwidth Design for failures Interface: non-posix New op: record appends (atomicity matters,
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung December 2003 ACM symposium on Operating systems principles Publisher: ACM Nov. 26, 2008 OUTLINE INTRODUCTION DESIGN OVERVIEW
More informationGoogle Disk Farm. Early days
Google Disk Farm Early days today CS 5204 Fall, 2007 2 Design Design factors Failures are common (built from inexpensive commodity components) Files large (multi-gb) mutation principally via appending
More informationDesigning Distributed Systems using Approximate Synchrony in Data Center Networks
Designing Distributed Systems using Approximate Synchrony in Data Center Networks Dan R. K. Ports Jialin Li Naveen Kr. Sharma Vincent Liu Arvind Krishnamurthy University of Washington CSE Today s most
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung SOSP 2003 presented by Kun Suo Outline GFS Background, Concepts and Key words Example of GFS Operations Some optimizations in
More informationChanging Requirements for Distributed File Systems in Cloud Storage
Changing Requirements for Distributed File Systems in Cloud Storage Wesley Leggette Cleversafe Presentation Agenda r About Cleversafe r Scalability, our core driver r Object storage as basis for filesystem
More informationOptimizing Software-Defined Storage for Flash Memory
Optimizing Software-Defined Storage for Flash Memory Avraham Meir Chief Scientist, Elastifile Santa Clara, CA 1 Agenda What is SDS? SDS Media Selection Flash-Native SDS Non-Flash SDS File system benefits
More informationEight Tips for Better Archives. Eight Ways Cloudian Object Storage Benefits Archiving with Veritas Enterprise Vault
Eight Tips for Better Email Archives Eight Ways Cloudian Object Storage Benefits Email Archiving with Veritas Enterprise Vault Most organizations now manage terabytes, if not petabytes, of corporate and
More informationExploiting Commutativity For Practical Fast Replication. Seo Jin Park and John Ousterhout
Exploiting Commutativity For Practical Fast Replication Seo Jin Park and John Ousterhout Overview Problem: consistent replication adds latency and throughput overheads Why? Replication happens after ordering
More informationLow-Latency Multi-Datacenter Databases using Replicated Commit
Low-Latency Multi-Datacenter Databases using Replicated Commit Hatem Mahmoud, Faisal Nawab, Alexander Pucher, Divyakant Agrawal, Amr El Abbadi UCSB Presented by Ashutosh Dhekne Main Contributions Reduce
More informationNew research on Key Technologies of unstructured data cloud storage
2017 International Conference on Computing, Communications and Automation(I3CA 2017) New research on Key Technologies of unstructured data cloud storage Songqi Peng, Rengkui Liua, *, Futian Wang State
More informationRecap. CSE 486/586 Distributed Systems Google Chubby Lock Service. Recap: First Requirement. Recap: Second Requirement. Recap: Strengthening P2
Recap CSE 486/586 Distributed Systems Google Chubby Lock Service Steve Ko Computer Sciences and Engineering University at Buffalo Paxos is a consensus algorithm. Proposers? Acceptors? Learners? A proposer
More informationSimpleChubby: a simple distributed lock service
SimpleChubby: a simple distributed lock service Jing Pu, Mingyu Gao, Hang Qu 1 Introduction We implement a distributed lock service called SimpleChubby similar to the original Google Chubby lock service[1].
More informationATOMIC COMMITMENT Or: How to Implement Distributed Transactions in Sharded Databases
ATOMIC COMMITMENT Or: How to Implement Distributed Transactions in Sharded Databases We talked about transactions and how to implement them in a single-node database. We ll now start looking into how to
More informationRecap. CSE 486/586 Distributed Systems Google Chubby Lock Service. Paxos Phase 2. Paxos Phase 1. Google Chubby. Paxos Phase 3 C 1
Recap CSE 486/586 Distributed Systems Google Chubby Lock Service Steve Ko Computer Sciences and Engineering University at Buffalo Paxos is a consensus algorithm. Proposers? Acceptors? Learners? A proposer
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google* 정학수, 최주영 1 Outline Introduction Design Overview System Interactions Master Operation Fault Tolerance and Diagnosis Conclusions
More informationcontext: massive systems
cutting the electric bill for internetscale systems Asfandyar Qureshi (MIT) Rick Weber (Akamai) Hari Balakrishnan (MIT) John Guttag (MIT) Bruce Maggs (Duke/Akamai) Éole @ flickr context: massive systems
More informationWindows Servers In Microsoft Azure
$6/Month Windows Servers In Microsoft Azure What I m Going Over 1. How inexpensive servers in Microsoft Azure are 2. How I get Windows servers for $6/month 3. Why Azure hosted servers are way better 4.
More informationI/O CANNOT BE IGNORED
LECTURE 13 I/O I/O CANNOT BE IGNORED Assume a program requires 100 seconds, 90 seconds for main memory, 10 seconds for I/O. Assume main memory access improves by ~10% per year and I/O remains the same.
More informationFuxiSort. Jiamang Wang, Yongjun Wu, Hua Cai, Zhipeng Tang, Zhiqiang Lv, Bin Lu, Yangyu Tao, Chao Li, Jingren Zhou, Hong Tang Alibaba Group Inc
Fuxi Jiamang Wang, Yongjun Wu, Hua Cai, Zhipeng Tang, Zhiqiang Lv, Bin Lu, Yangyu Tao, Chao Li, Jingren Zhou, Hong Tang Alibaba Group Inc {jiamang.wang, yongjun.wyj, hua.caihua, zhipeng.tzp, zhiqiang.lv,
More informationSCALABLE CONSISTENCY AND TRANSACTION MODELS
Data Management in the Cloud SCALABLE CONSISTENCY AND TRANSACTION MODELS 69 Brewer s Conjecture Three properties that are desirable and expected from realworld shared-data systems C: data consistency A:
More informationClouds are complex so they fail. Cloud File System Security and Dependability with SafeCloud-FS
Cloud File System Security and ependability with SafeCloud-FS Miguel P. Correia KTH Stockholm June 2017 Joint work with Alysson Bessani, B. Quaresma, F. André, P. Sousa, R. Mendes, T. Oliveira, N. Neves,
More informationOasis: An Active Storage Framework for Object Storage Platform
Oasis: An Active Storage Framework for Object Storage Platform Yulai Xie 1, Dan Feng 1, Darrell D. E. Long 2, Yan Li 2 1 School of Computer, Huazhong University of Science and Technology Wuhan National
More informationDecentralized Distributed Storage System for Big Data
Decentralized Distributed Storage System for Big Presenter: Wei Xie -Intensive Scalable Computing Laboratory(DISCL) Computer Science Department Texas Tech University Outline Trends in Big and Cloud Storage
More informationLecture 21: Reliable, High Performance Storage. CSC 469H1F Fall 2006 Angela Demke Brown
Lecture 21: Reliable, High Performance Storage CSC 469H1F Fall 2006 Angela Demke Brown 1 Review We ve looked at fault tolerance via server replication Continue operating with up to f failures Recovery
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff and Shun Tak Leung Google* Shivesh Kumar Sharma fl4164@wayne.edu Fall 2015 004395771 Overview Google file system is a scalable distributed file system
More informationCloud Computing and Hadoop Distributed File System. UCSB CS170, Spring 2018
Cloud Computing and Hadoop Distributed File System UCSB CS70, Spring 08 Cluster Computing Motivations Large-scale data processing on clusters Scan 000 TB on node @ 00 MB/s = days Scan on 000-node cluster
More informationEECS 498 Introduction to Distributed Systems
EECS 498 Introduction to Distributed Systems Fall 2017 Harsha V. Madhyastha November 27, 2017 EECS 498 Lecture 19 2 Windows Azure Storage Focus on storage within a data center Goals: Durability Strong
More informationECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective
ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective Part II: Data Center Software Architecture: Topic 1: Distributed File Systems GFS (The Google File System) 1 Filesystems
More informationKernel Level Speculative DSM
Motivation Main interest is performance, fault-tolerance, and correctness of distributed systems Present our ideas in the context of a DSM system We are developing tools that Improve performance Address
More informationRepair Pipelining for Erasure-Coded Storage
Repair Pipelining for Erasure-Coded Storage Runhui Li, Xiaolu Li, Patrick P. C. Lee, Qun Huang The Chinese University of Hong Kong USENIX ATC 2017 1 Introduction Fault tolerance for distributed storage
More information25 Best Practice Tips for architecting Amazon VPC
25 Best Practice Tips for architecting Amazon VPC 25 Best Practice Tips for architecting Amazon VPC Amazon VPC is one of the most important feature introduced by AWS. We have been using AWS from 2008 and
More information! Design constraints. " Component failures are the norm. " Files are huge by traditional standards. ! POSIX-like
Cloud background Google File System! Warehouse scale systems " 10K-100K nodes " 50MW (1 MW = 1,000 houses) " Power efficient! Located near cheap power! Passive cooling! Power Usage Effectiveness = Total
More informationLocally repairable codes
Locally repairable codes for large-scale storage Presented by Anwitaman Datta Nanyang Technological University, Singapore Joint work with Frédérique Oggier & Lluís Pàmies i Juárez @ IISc 2012 & NetApp
More informationGoogle File System. By Dinesh Amatya
Google File System By Dinesh Amatya Google File System (GFS) Sanjay Ghemawat, Howard Gobioff, Shun-Tak Leung designed and implemented to meet rapidly growing demand of Google's data processing need a scalable
More informationA Reliable Broadcast System
A Reliable Broadcast System Yuchen Dai, Xiayi Huang, Diansan Zhou Department of Computer Sciences and Engineering Santa Clara University December 10 2013 Table of Contents 2 Introduction......3 2.1 Objective...3
More informationThe New Economics of Cloud Storage
The New Economics of Cloud Storage How Wasabi Hot Cloud Storage Compares With the Economics of Amazon Web Services, Google Cloud and Microsoft Azure Executive Overview Wasabi is fundamentally transforming
More informationManaging Array of SSDs When the Storage Device is No Longer the Performance Bottleneck
Managing Array of Ds When the torage Device is No Longer the Performance Bottleneck Byung. Kim, Jaeho Kim, am H. Noh UNIT (Ulsan National Institute of cience & Technology) Outline Motivation & Observation
More informationData Modeling and Databases Ch 14: Data Replication. Gustavo Alonso, Ce Zhang Systems Group Department of Computer Science ETH Zürich
Data Modeling and Databases Ch 14: Data Replication Gustavo Alonso, Ce Zhang Systems Group Department of Computer Science ETH Zürich Database Replication What is database replication The advantages of
More informationIntroduction to MapReduce
Basics of Cloud Computing Lecture 4 Introduction to MapReduce Satish Srirama Some material adapted from slides by Jimmy Lin, Christophe Bisciglia, Aaron Kimball, & Sierra Michels-Slettvet, Google Distributed
More informationDatacenter replication solution with quasardb
Datacenter replication solution with quasardb Technical positioning paper April 2017 Release v1.3 www.quasardb.net Contact: sales@quasardb.net Quasardb A datacenter survival guide quasardb INTRODUCTION
More informationMDCC MULTI DATA CENTER CONSISTENCY. amplab. Tim Kraska, Gene Pang, Michael Franklin, Samuel Madden, Alan Fekete
MDCC MULTI DATA CENTER CONSISTENCY Tim Kraska, Gene Pang, Michael Franklin, Samuel Madden, Alan Fekete gpang@cs.berkeley.edu amplab MOTIVATION 2 3 June 2, 200: Rackspace power outage of approximately 0
More informationThe Google File System
The Google File System By Ghemawat, Gobioff and Leung Outline Overview Assumption Design of GFS System Interactions Master Operations Fault Tolerance Measurements Overview GFS: Scalable distributed file
More informationStorage Integration with Host-based Write-back Caching
Storage Integration with Host-based Write-back Caching Andy Banta @andybanta NetApp SolidFire Santa Clara, CA 1 Agenda Patented information How virtual machines use storage Caching methods And who can
More informationGeorgia Institute of Technology ECE6102 4/20/2009 David Colvin, Jimmy Vuong
Georgia Institute of Technology ECE6102 4/20/2009 David Colvin, Jimmy Vuong Relatively recent; still applicable today GFS: Google s storage platform for the generation and processing of data used by services
More informationCSE 124: Networked Services Lecture-17
Fall 2010 CSE 124: Networked Services Lecture-17 Instructor: B. S. Manoj, Ph.D http://cseweb.ucsd.edu/classes/fa10/cse124 11/30/2010 CSE 124 Networked Services Fall 2010 1 Updates PlanetLab experiments
More informationNPTEL Course Jan K. Gopinath Indian Institute of Science
Storage Systems NPTEL Course Jan 2012 (Lecture 39) K. Gopinath Indian Institute of Science Google File System Non-Posix scalable distr file system for large distr dataintensive applications performance,
More informationA Practical Scalable Distributed B-Tree
A Practical Scalable Distributed B-Tree CS 848 Paper Presentation Marcos K. Aguilera, Wojciech Golab, Mehul A. Shah PVLDB 08 March 8, 2010 Presenter: Evguenia (Elmi) Eflov Presentation Outline 1 Background
More informationTAPIR. By Irene Zhang, Naveen Sharma, Adriana Szekeres, Arvind Krishnamurthy, and Dan Ports Presented by Todd Charlton
TAPIR By Irene Zhang, Naveen Sharma, Adriana Szekeres, Arvind Krishnamurthy, and Dan Ports Presented by Todd Charlton Outline Problem Space Inconsistent Replication TAPIR Evaluation Conclusion Problem
More informationGFS: The Google File System
GFS: The Google File System Brad Karp UCL Computer Science CS GZ03 / M030 24 th October 2014 Motivating Application: Google Crawl the whole web Store it all on one big disk Process users searches on one
More informationPerformance Models of Access Latency in Cloud Storage Systems
Performance Models of Access Latency in Cloud Storage Systems Qiqi Shuai Email: qqshuai@eee.hku.hk Victor O.K. Li, Fellow, IEEE Email: vli@eee.hku.hk Yixuan Zhu Email: yxzhu@eee.hku.hk Abstract Access
More informationYour Data is in the Cloud: Who Exactly is Looking After It?
Your Data is in the Cloud: Who Exactly is Looking After It? P Vijay Kumar Dept of Electrical Communication Engineering Indian Institute of Science IISc Open Day March 4, 2017 1/33 Your Data is in the Cloud:
More informationHedvig as backup target for Veeam
Hedvig as backup target for Veeam Solution Whitepaper Version 1.0 April 2018 Table of contents Executive overview... 3 Introduction... 3 Solution components... 4 Hedvig... 4 Hedvig Virtual Disk (vdisk)...
More informationA Comprehensive Analytical Performance Model of DRAM Caches
A Comprehensive Analytical Performance Model of DRAM Caches Authors: Nagendra Gulur *, Mahesh Mehendale *, and R Govindarajan + Presented by: Sreepathi Pai * Texas Instruments, + Indian Institute of Science
More informationBigtable. A Distributed Storage System for Structured Data. Presenter: Yunming Zhang Conglong Li. Saturday, September 21, 13
Bigtable A Distributed Storage System for Structured Data Presenter: Yunming Zhang Conglong Li References SOCC 2010 Key Note Slides Jeff Dean Google Introduction to Distributed Computing, Winter 2008 University
More informationDistributed DNS Name server Backed by Raft
Distributed DNS Name server Backed by Raft Emre Orbay Gabbi Fisher December 13, 2017 Abstract We implement a asynchronous distributed DNS server backed by Raft. Our DNS server system supports large cluster
More informationMillWheel:Fault Tolerant Stream Processing at Internet Scale. By FAN Junbo
MillWheel:Fault Tolerant Stream Processing at Internet Scale By FAN Junbo Introduction MillWheel is a low latency data processing framework designed by Google at Internet scale. Motived by Google Zeitgeist
More informationCeph: A Scalable, High-Performance Distributed File System
Ceph: A Scalable, High-Performance Distributed File System S. A. Weil, S. A. Brandt, E. L. Miller, D. D. E. Long Presented by Philip Snowberger Department of Computer Science and Engineering University
More informationHigh Performance Transactions in Deuteronomy
High Performance Transactions in Deuteronomy Justin Levandoski, David Lomet, Sudipta Sengupta, Ryan Stutsman, and Rui Wang Microsoft Research Overview Deuteronomy: componentized DB stack Separates transaction,
More informationData Centers. Tom Anderson
Data Centers Tom Anderson Transport Clarification RPC messages can be arbitrary size Ex: ok to send a tree or a hash table Can require more than one packet sent/received We assume messages can be dropped,
More informationThe Google File System
October 13, 2010 Based on: S. Ghemawat, H. Gobioff, and S.-T. Leung: The Google file system, in Proceedings ACM SOSP 2003, Lake George, NY, USA, October 2003. 1 Assumptions Interface Architecture Single
More informationKubernetes Integration with Virtuozzo Storage
Kubernetes Integration with Virtuozzo Storage A Technical OCTOBER, 2017 2017 Virtuozzo. All rights reserved. 1 Application Container Storage Application containers appear to be the perfect tool for supporting
More informationBuilding Consistent Transactions with Inconsistent Replication
DB Reading Group Fall 2015 slides by Dana Van Aken Building Consistent Transactions with Inconsistent Replication Irene Zhang, Naveen Kr. Sharma, Adriana Szekeres, Arvind Krishnamurthy, Dan R. K. Ports
More informationOperating Systems. Lecture File system implementation. Master of Computer Science PUF - Hồ Chí Minh 2016/2017
Operating Systems Lecture 7.2 - File system implementation Adrien Krähenbühl Master of Computer Science PUF - Hồ Chí Minh 2016/2017 Design FAT or indexed allocation? UFS, FFS & Ext2 Journaling with Ext3
More informationElastic Cloud Storage (ECS)
Elastic Cloud Storage (ECS) Version 3.1 Administration Guide 302-003-863 02 Copyright 2013-2017 Dell Inc. or its subsidiaries. All rights reserved. Published September 2017 Dell believes the information
More informationRouting in Delay Tolerant Networks (2)
Routing in Delay Tolerant Networks (2) Primary Reference: E. P. C. Jones, L. Li and P. A. S. Ward, Practical Routing in Delay-Tolerant Networks, SIGCOMM 05, Workshop on DTN, August 22-26, 2005, Philadelphia,
More informationVirtual Storage Tier and Beyond
Virtual Storage Tier and Beyond Manish Agarwal Sr. Product Manager, NetApp Santa Clara, CA 1 Agenda Trends Other Storage Trends and Flash 5 Min Rule Issues for Flash Dedupe and Flash Caching Architectural
More informationTripS: Automated Multi-tiered Data Placement in a Geo-distributed Cloud Environment
TripS: Automated Multi-tiered Data Placement in a Geo-distributed Cloud Environment Kwangsung Oh, Abhishek Chandra, and Jon Weissman Department of Computer Science and Engineering University of Minnesota
More informationDeukyeon Hwang UNIST. Wook-Hee Kim UNIST. Beomseok Nam UNIST. Hanyang Univ.
Deukyeon Hwang UNIST Wook-Hee Kim UNIST Youjip Won Hanyang Univ. Beomseok Nam UNIST Fast but Asymmetric Access Latency Non-Volatility Byte-Addressability Large Capacity CPU Caches (Volatile) Persistent
More informationConsolidating Concurrency Control and Consensus for Commits under Conflicts
Consolidating Concurrency Control and Consensus for Commits under Conflicts Shuai Mu and Lamont Nelson, New York University; Wyatt Lloyd, University of Southern California; Jinyang Li, New York University
More informationDistributed Systems. Tutorial 9 Windows Azure Storage
Distributed Systems Tutorial 9 Windows Azure Storage written by Alex Libov Based on SOSP 2011 presentation winter semester, 2011-2012 Windows Azure Storage (WAS) A scalable cloud storage system In production
More informationOutline. Spanner Mo/va/on. Tom Anderson
Spanner Mo/va/on Tom Anderson Outline Last week: Chubby: coordina/on service BigTable: scalable storage of structured data GFS: large- scale storage for bulk data Today/Friday: Lessons from GFS/BigTable
More informationgoals monitoring, fault tolerance, auto-recovery (thousands of low-cost machines) handle appends efficiently (no random writes & sequential reads)
Google File System goals monitoring, fault tolerance, auto-recovery (thousands of low-cost machines) focus on multi-gb files handle appends efficiently (no random writes & sequential reads) co-design GFS
More informationLive Virtual Machine Migration with Efficient Working Set Prediction
2011 International Conference on Network and Electronics Engineering IPCSIT vol.11 (2011) (2011) IACSIT Press, Singapore Live Virtual Machine Migration with Efficient Working Set Prediction Ei Phyu Zaw
More informationMapReduce. U of Toronto, 2014
MapReduce U of Toronto, 2014 http://www.google.org/flutrends/ca/ (2012) Average Searches Per Day: 5,134,000,000 2 Motivation Process lots of data Google processed about 24 petabytes of data per day in
More informationPervasive PSQL Vx Server Licensing
Pervasive PSQL Vx Server Licensing Overview The Pervasive PSQL Vx Server edition is designed for highly virtualized environments with support for enterprise hypervisor features including live application
More informationCORFU: A Shared Log Design for Flash Clusters
CORFU: A Shared Log Design for Flash Clusters Authors: Mahesh Balakrishnan, Dahlia Malkhi, Vijayan Prabhakaran, Ted Wobber, Michael Wei, John D. Davis EECS 591 11/7/18 Presented by Evan Agattas and Fanzhong
More information