End-to-End Data Integrity in the Intel/EMC/HDF Group Exascale IO DOE Fast Forward Project
|
|
- Aubrey Robertson
- 5 years ago
- Views:
Transcription
1 End-to-End Data Integrity in the Intel/EMC/HDF Group Exascale IO DOE Fast Forward Project As presented by John Bent, EMC and Quincey Koziol, The HDF Group
2 Truly End-to-End App provides checksum buffer to HDF5 Input checksum on write, return value on read Optional. Can be disabled or ask HDF5 to do it. HDF5 passes to IOD IOD does necessary recomputation when unaligned Buck stops at IOD. DAOS doesn t (yet) participate IOD checksums stored as regular DAOS data DAOS and zfs and lustre all do checksumming as well... Note: We prevent silent data corruption only We don t (yet) repair
3 Checksum Support in the API s Every IO buffer will have a checksum A checksum function is provided to upper layers E.g. iod_checksum_t iod_checksum(buffer)
4 Data Integrity in the Stack during Writes 1. HDF5 API allows checksums to be optionally passed Application 2. If app doesn t pass checksum, it will be added somewhere in the HDF5/IOD VOL layers H5Dwrite(data,(checksum)) HDF / IOD VOL iod_obj_write(data,checksum) 3. The function shipper just does a passthrough 4. IOD does two writes into DAOS. IOD can actually create a DAOS (virtual) shard that is optimized for small iops to store checksums and a DAOS (virtual) shard optimized for bandwidth to store data. 5. DAOS stores the data (which is actually metadata and data for the above layer) Function Shipper iod_obj_write(data,checksum) IOD daos_shard_write(data) daos_shard_write(checksum) DAOS
5 Data Integrity in the Stack during Reads 1. The read request goes down the stack and gets to IOD 2. IOD reads the data and checksum from DAOS. May require multiple reads of multiple buffers and their verification and a recomputation of a new checksum if unaligned. Being careful to avoid race conditions. Application H5Dread(data,(&checksum?)) HDF / IOD VOL iod_obj_read(data,&checksum) Function Shipper iod_obj_read(data,checksum) IOD 3. IOD returns the data and the checksum up the stack 1. Hints can disable this daos_shard_read(data) daos_shard_read(checksum) DAOS
6 An HDF5 Dataset is stored in a logical IOD Array and nicely striped across a set of DAOS shards. Each cell has its own checksum. DAOS Storage Target DAOS Shard
7 An HDF5 Dataset is stored in a logical IOD Array and nicely striped across a set of DAOS shards. Each cell has its own checksum. An aligned full cell read is easy! Just return the cell and its checksum. DAOS Storage Target DAOS Shard
8 An HDF5 Dataset is stored in a logical IOD Array and nicely striped across a set of DAOS shards. Each cell has its own checksum. An unaligned read is hard! Imagine reading this bright pink rectangle. Many checksum computations are required and race conditions must be carefully avoided. [Hints can disable if performance is paramount over integrity.] DAOS Storage Target DAOS Shard
9 Race Conditions, Contiguous Blocks Imagine a read straddling checksummed blocks {cksum1} {ret_cksum} {cksum2} Create ret_cksum, then verify existing cksums, then copy KEY: Create must happen first If verify and then create, corruption can occur between KEY: Create ret_cksum from existing data not from copy If create from copy, copy might already be corrupted
10 Race Conditions, Non-Contiguous Imagine a read straddling checksummed non-contiguous blocks {cksum1} {tmp1} {tmp2} {cksum2} Creating ret_cksum cannot be done in one operation Especially if regions come from more than one storage node In this case, each region must be cksum d, then copied. Then create ret_cksum on return buffer. The cksum each copied region and compare to temporary cksums on the source regions. Then, finally, verify the original cksums on the blocks.
11 Non-contiguous reads {ret_cksum} {tmp1.1} {tmp2.1} {cksum1} {tmp1} {tmp2} {cksum2} First create tmp1 and tmp2 cksums Then copy regions Then create ret_cksum from copy Then create tmp1.1 and tmp2.1 and compare to tmp1 and tmp2 Then verify cksum1 and cksum2
12 Why not Verify First? {ret_cksum} {tmp1.1} {tmp2.1} {cksum1} First verify cksum1 and cksum2 Then create tmp1 and tmp2 cksums Then copy regions Then create ret_cksum from copy Then create tmp1.1 and tmp2.1 and compare to tmp1 and tmp2 {tmp1} {tmp2} {cksum2}
13 What about writes? Just like reads but in reverse
14 Read Pseudo-code iod_obj_read(offset=o,length=l,checksum=&c,buffer=&b) { regions = find_all_data(o,l) foreach region and its checksum (R,V) in regions # the checksum will be on disk for whole regions # or recomputed if a partial region checksums[r] = V # save the checksum copy R into B *C = checksum(b) # checksum entire output buffer foreach region R in regions # verify copied region within buffer matches original's checksum checksum = checksum(r within B) assert(checksum == checksums[r]) }
15 Three IOD Object Types Blobs Just as has been described Arrays When stored, they are unrolled into a blob KV Stores Store checksum(s) as a header in value Other metadata may be stored here as well such as value length
16 Storing Data In ION, we will store the data in traditional PLFS style logfiles Pro: fast writes Con: large amount of IOD/PLFS index metadata but it s short-lived In DAOS, we will store the data in flattened view as round-robined stripes across shards Pro: minimized IOD metadata Con: potentially slower migrate due to need to scatter/gather but it s on fast interconnect
17 Storing Checksums In ION, we will store the checksum in a new field in the PLFS index entry for each range Pro: very easy to implement Con: can t do pattern compression on PLFS index In DAOS, we will store the checksum in a virtual checksum shard corresponding to the object of the same number in a virtual data shard
18 Storing IOD Objects Onto DAOS Split each DAOS shard into four virtual shards Metadata (00) Checksum (01) Data (10) Reserved (11) [in future for DAOS HA?] An IOD object ID, OID, is 62 bits To read its metadata, read DAOS object {00}{OID} To read its checksum, read DAOS object {01}{OID} To read its data, read DAOS object {10}{OID}
19 IOD Metadata Found by getting list of shards from container Hash(OID) % shard list to find target shard Read {00}{OID} on that shard to get metadata Metadata is very small A list of shards across which this object is striped Checksum unit size Tunable via hints or explicit parameters or IOD Stripe size observation of usage while in ION Multiple of the checksum unit size The last offset of the object The dimensionality info for array objects
20 IOD Data Data for IOD object OID is striped in a roundrobin fashion in object {10}{OID} across a set of shards Since DAOS is very good at sparse objects and can flatten overwrites nicely with transactions, we place all data at the same physical offset as the logical target offset (this reduces our metadata)
21 IOD Checksums The data for block B of object OID is stored at {10}{OID} in the appropriate shard given the stripe size and list of shards for OID (as explained in previous slide) Therefore, the checksum for block B is stored at {01}{OID} in that same shard. Even though DAOS is good at sparse, we might want to avoid very small IO s if we put checksums at the same offset as the block they describe. Therefore, we may instead do an array of checksums for each stripe squished together in the front of the corresponding stripe of the checksum shard
22 Storing OID 3 on DAOS Shards 3 % 2 = 1. Metadata at shard 1 {00}{3} Data is striped across objects {10}{3} on each shard, starting at shard 1 Checksums for each stripe at corresponding location in objects {01}{3} {meta} {cksums} {data} {empty}
23 Zooming in on Checksum Block {meta} {empty} {cksums} {data}
24 Zooming in on Checksum Block This shows one data block and one checksum block holding one single checksum. But actually each data block is split into multiple checksum units. Each checksum could be at same offset as its checksum unit. But this is too sparse (e.g. 64 bits every MB). Instead create an array of checksums for each checksum unit at the front of the checksum block.
25 More info Optimizations about sub-chunking checksum regions are available in this white paper by Andreas Dilger: Integrity pdf ( )
26 HDF5 Metadata End-to-End Integrity
27 HDF5 Metadata End-to-End Integrity
28 HDF5 Metadata End-to-End Integrity
29 HDF5 Metadata End-to-End Integrity
30 HDF5 Metadata End-to-End Integrity
31 HDF5 Metadata End-to-End Integrity
32 HDF5 Metadata End-to-End Integrity
33 HDF5 Metadata End-to-End Integrity
34 HDF5 Metadata End-to-End Integrity
35 HDF5 Metadata End-to-End Integrity
36 HDF5 Metadata End-to-End Integrity
37 HDF5 Metadata End-to-End Integrity
38 HDF5 Metadata End-to-End Integrity
39 HDF5 Metadata End-to-End Integrity
40 HDF5 Metadata End-to-End Integrity
41 HDF5 Raw Data End-to-End Integrity
FastForward I/O and Storage: IOD M5 Demonstration (5.2, 5.3, 5.9, 5.10)
FastForward I/O and Storage: IOD M5 Demonstration (5.2, 5.3, 5.9, 5.10) 1 EMC September, 2013 John Bent john.bent@emc.com Sorin Faibish faibish_sorin@emc.com Xuezhao Liu xuezhao.liu@emc.com Harriet Qiu
More information8.5 End-to-End Demonstration Exascale Fast Forward Storage Team June 30 th, 2014
8.5 End-to-End Demonstration Exascale Fast Forward Storage Team June 30 th, 2014 NOTICE: THIS MANUSCRIPT HAS BEEN AUTHORED BY INTEL, THE HDF GROUP, AND EMC UNDER INTEL S SUBCONTRACT WITH LAWRENCE LIVERMORE
More informationEFF-IO M7.5 Demo. Semantic Migration of Multi-dimensional Arrays
EFF-IO M7.5 Demo Semantic Migration of Multi-dimensional Arrays John Bent, Sorin Faibish, Xuezhao Liu, Harriet Qui, Haiying Tang, Jerry Tirrell, Jingwang Zhang, Kelly Zhang, Zhenhua Zhang NOTICE: THIS
More informationCaching and Buffering in HDF5
Caching and Buffering in HDF5 September 9, 2008 SPEEDUP Workshop - HDF5 Tutorial 1 Software stack Life cycle: What happens to data when it is transferred from application buffer to HDF5 file and from HDF5
More informationHigh Level Design IOD KV Store FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O
Date: January 10, 2013 High Level Design IOD KV Store FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O LLNS Subcontract No. Subcontractor Name Subcontractor Address B599860
More informationA Plugin for HDF5 using PLFS for Improved I/O Performance and Semantic Analysis
2012 SC Companion: High Performance Computing, Networking Storage and Analysis A for HDF5 using PLFS for Improved I/O Performance and Semantic Analysis Kshitij Mehta, John Bent, Aaron Torres, Gary Grider,
More informationDRAFT. HDF5 Data Flow Pipeline for H5Dread. 1 Introduction. 2 Examples
This document describes the HDF5 library s data movement and processing activities when H5Dread is called for a dataset with chunked storage. The document provides an overview of how memory management,
More information<Insert Picture Here> End-to-end Data Integrity for NFS
End-to-end Data Integrity for NFS Chuck Lever Consulting Member of Technical Staff Today s Discussion What is end-to-end data integrity? T10 PI overview Adapting
More informationFile Open, Close, and Flush Performance Issues in HDF5 Scot Breitenfeld John Mainzer Richard Warren 02/19/18
File Open, Close, and Flush Performance Issues in HDF5 Scot Breitenfeld John Mainzer Richard Warren 02/19/18 1 Introduction Historically, the parallel version of the HDF5 library has suffered from performance
More informationUK LUG 10 th July Lustre at Exascale. Eric Barton. CTO Whamcloud, Inc Whamcloud, Inc.
UK LUG 10 th July 2012 Lustre at Exascale Eric Barton CTO Whamcloud, Inc. eeb@whamcloud.com Agenda Exascale I/O requirements Exascale I/O model 3 Lustre at Exascale - UK LUG 10th July 2012 Exascale I/O
More informationExtreme I/O Scaling with HDF5
Extreme I/O Scaling with HDF5 Quincey Koziol Director of Core Software Development and HPC The HDF Group koziol@hdfgroup.org July 15, 2012 XSEDE 12 - Extreme Scaling Workshop 1 Outline Brief overview of
More informationRFC: HDF5 File Space Management: Paged Aggregation
RFC: HDF5 File Space Management: Paged Aggregation Vailin Choi Quincey Koziol John Mainzer The current HDF5 file space allocation accumulates small pieces of metadata and raw data in aggregator blocks.
More informationHDF5 I/O Performance. HDF and HDF-EOS Workshop VI December 5, 2002
HDF5 I/O Performance HDF and HDF-EOS Workshop VI December 5, 2002 1 Goal of this talk Give an overview of the HDF5 Library tuning knobs for sequential and parallel performance 2 Challenging task HDF5 Library
More informationDAOS and Friends: A Proposal for an Exascale Storage System
DAOS and Friends: A Proposal for an Exascale Storage System Jay Lofstead, Ivo Jimenez, Carlos Maltzahn, Quincey Koziol, John Bent, Eric Barton Sandia National Laboratories gflofst@sandia.gov University
More informationCS370: System Architecture & Software [Fall 2014] Dept. Of Computer Science, Colorado State University
CS 370: SYSTEM ARCHITECTURE & SOFTWARE [MASS STORAGE] Frequently asked questions from the previous class survey Shrideep Pallickara Computer Science Colorado State University L29.1 L29.2 Topics covered
More informationThe What, Why and How of the Pure Storage Enterprise Flash Array. Ethan L. Miller (and a cast of dozens at Pure Storage)
The What, Why and How of the Pure Storage Enterprise Flash Array Ethan L. Miller (and a cast of dozens at Pure Storage) Enterprise storage: $30B market built on disk Key players: EMC, NetApp, HP, etc.
More informationCSE 124: Networked Services Fall 2009 Lecture-19
CSE 124: Networked Services Fall 2009 Lecture-19 Instructor: B. S. Manoj, Ph.D http://cseweb.ucsd.edu/classes/fa09/cse124 Some of these slides are adapted from various sources/individuals including but
More informationFast Forward I/O & Storage
Fast Forward I/O & Storage Eric Barton Lead Architect 1 Department of Energy - Fast Forward Challenge FastForward RFP provided US Government funding for exascale research and development Sponsored by 7
More informationBigTable: A Distributed Storage System for Structured Data (2006) Slides adapted by Tyler Davis
BigTable: A Distributed Storage System for Structured Data (2006) Slides adapted by Tyler Davis Motivation Lots of (semi-)structured data at Google URLs: Contents, crawl metadata, links, anchors, pagerank,
More informationCSE 124: Networked Services Lecture-16
Fall 2010 CSE 124: Networked Services Lecture-16 Instructor: B. S. Manoj, Ph.D http://cseweb.ucsd.edu/classes/fa10/cse124 11/23/2010 CSE 124 Networked Services Fall 2010 1 Updates PlanetLab experiments
More informationDesign Document (Historical) HDF5 Dynamic Data Structure Support FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O
Date: July 24, 2013 Design Document (Historical) HDF5 Dynamic Data Structure Support FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O LLNS Subcontract No. Subcontractor
More informationA Private Heap for HDF5 Quincey Koziol Jan. 15, 2007
A Private Heap for HDF5 Quincey Koziol Jan. 15, 2007 Background The HDF5 library currently stores variable-sized data in two different data structures in its files. Variable-sized metadata (currently only
More informationOperating Systems. Lecture File system implementation. Master of Computer Science PUF - Hồ Chí Minh 2016/2017
Operating Systems Lecture 7.2 - File system implementation Adrien Krähenbühl Master of Computer Science PUF - Hồ Chí Minh 2016/2017 Design FAT or indexed allocation? UFS, FFS & Ext2 Journaling with Ext3
More informationlibhio: Optimizing IO on Cray XC Systems With DataWarp
libhio: Optimizing IO on Cray XC Systems With DataWarp May 9, 2017 Nathan Hjelm Cray Users Group May 9, 2017 Los Alamos National Laboratory LA-UR-17-23841 5/8/2017 1 Outline Background HIO Design Functionality
More informationLock Ahead: Shared File Performance Improvements
Lock Ahead: Shared File Performance Improvements Patrick Farrell Cray Lustre Developer Steve Woods Senior Storage Architect woods@cray.com September 2016 9/12/2016 Copyright 2015 Cray Inc 1 Agenda Shared
More informationWhat NetCDF users should know about HDF5?
What NetCDF users should know about HDF5? Elena Pourmal The HDF Group July 20, 2007 7/23/07 1 Outline The HDF Group and HDF software HDF5 Data Model Using HDF5 tools to work with NetCDF-4 programs files
More informationThe Google File System
The Google File System By Ghemawat, Gobioff and Leung Outline Overview Assumption Design of GFS System Interactions Master Operations Fault Tolerance Measurements Overview GFS: Scalable distributed file
More informationh5perf_serial, a Serial File System Benchmarking Tool
h5perf_serial, a Serial File System Benchmarking Tool The HDF Group April, 2009 HDF5 users have reported the need to perform serial benchmarking on systems without an MPI environment. The parallel benchmarking
More informationZFS STORAGE POOL LAYOUT. Storage and Servers Driven by Open Source.
ZFS STORAGE POOL LAYOUT Storage and Servers Driven by Open Source marketing@ixsystems.com CONTENTS 1 Introduction and Executive Summary 2 Striped vdev 3 Mirrored vdev 4 RAIDZ vdev 5 Examples by Workload
More informationCS370: Operating Systems [Spring 2017] Dept. Of Computer Science, Colorado State University
CS 370: OPERATING SYSTEMS [DISK SCHEDULING ALGORITHMS] Shrideep Pallickara Computer Science Colorado State University Frequently asked questions from the previous class survey Can a UNIX file span over
More informationBigtable. A Distributed Storage System for Structured Data. Presenter: Yunming Zhang Conglong Li. Saturday, September 21, 13
Bigtable A Distributed Storage System for Structured Data Presenter: Yunming Zhang Conglong Li References SOCC 2010 Key Note Slides Jeff Dean Google Introduction to Distributed Computing, Winter 2008 University
More informationOverview of High Performance Input/Output on LRZ HPC systems. Christoph Biardzki Richard Patra Reinhold Bader
Overview of High Performance Input/Output on LRZ HPC systems Christoph Biardzki Richard Patra Reinhold Bader Agenda Choosing the right file system Storage subsystems at LRZ Introduction to parallel file
More informationECSS Project: Prof. Bodony: CFD, Aeroacoustics
ECSS Project: Prof. Bodony: CFD, Aeroacoustics Robert McLay The Texas Advanced Computing Center June 19, 2012 ECSS Project: Bodony Aeroacoustics Program Program s name is RocfloCM It is mixture of Fortran
More informationLustre* - Fast Forward to Exascale High Performance Data Division. Eric Barton 18th April, 2013
Lustre* - Fast Forward to Exascale High Performance Data Division Eric Barton 18th April, 2013 DOE Fast Forward IO and Storage Exascale R&D sponsored by 7 leading US national labs Solutions to currently
More informationBest Practices in Designing Cloud Storage based Archival solution Sreenidhi Iyangar & Jim Rice EMC Corporation
Best Practices in Designing Cloud Storage based Archival solution Sreenidhi Iyangar & Jim Rice EMC Corporation Abstract Cloud storage facilitates the use case of digital archiving for long periods of time
More informationCS 390 Chapter 8 Homework Solutions
CS 390 Chapter 8 Homework Solutions 8.3 Why are page sizes always... Page sizes that are a power of two make it computationally fast for the kernel to determine the page number and offset of a logical
More informationAn Evolutionary Path to Object Storage Access
An Evolutionary Path to Object Storage Access David Goodell +, Seong Jo (Shawn) Kim*, Robert Latham +, Mahmut Kandemir*, and Robert Ross + *Pennsylvania State University + Argonne National Laboratory Outline
More informationCS 318 Principles of Operating Systems
CS 318 Principles of Operating Systems Fall 2018 Lecture 16: Advanced File Systems Ryan Huang Slides adapted from Andrea Arpaci-Dusseau s lecture 11/6/18 CS 318 Lecture 16 Advanced File Systems 2 11/6/18
More informationWhite Paper. Nexenta Replicast
White Paper Nexenta Replicast By Caitlin Bestler, September 2013 Table of Contents Overview... 3 Nexenta Replicast Description... 3 Send Once, Receive Many... 4 Distributed Storage Basics... 7 Nexenta
More informationIntroduction to High Performance Parallel I/O
Introduction to High Performance Parallel I/O Richard Gerber Deputy Group Lead NERSC User Services August 30, 2013-1- Some slides from Katie Antypas I/O Needs Getting Bigger All the Time I/O needs growing
More informationDisclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme
STO1926BU A Day in the Life of a VSAN I/O Diving in to the I/O Flow of vsan John Nicholson (@lost_signal) Pete Koehler (@vmpete) VMworld 2017 Content: Not for publication #VMworld #STO1926BU Disclaimer
More informationVirtualization Overview NSRC
Virtualization Overview NSRC Terminology Virtualization: dividing available resources into smaller independent units Emulation: using software to simulate hardware which you do not have The two often come
More informationToday: Coda, xfs. Case Study: Coda File System. Brief overview of other file systems. xfs Log structured file systems HDFS Object Storage Systems
Today: Coda, xfs Case Study: Coda File System Brief overview of other file systems xfs Log structured file systems HDFS Object Storage Systems Lecture 20, page 1 Coda Overview DFS designed for mobile clients
More informationAnnouncements. Persistence: Crash Consistency
Announcements P4 graded: In Learn@UW by end of day P5: Available - File systems Can work on both parts with project partner Fill out form BEFORE tomorrow (WED) morning for match Watch videos; discussion
More informationLustre Clustered Meta-Data (CMD) Huang Hua Andreas Dilger Lustre Group, Sun Microsystems
Lustre Clustered Meta-Data (CMD) Huang Hua H.Huang@Sun.Com Andreas Dilger adilger@sun.com Lustre Group, Sun Microsystems 1 Agenda What is CMD? How does it work? What are FIDs? CMD features CMD tricks Upcoming
More informationFast Forward Storage & I/O. Jeff Layton (Eric Barton)
Fast Forward & I/O Jeff Layton (Eric Barton) DOE Fast Forward IO and Exascale R&D sponsored by 7 leading US national labs Solutions to currently intractable problems of Exascale required to meet the 2020
More informationIntroduction to carving File fragmentation Object validation Carving methods Conclusion
Simson L. Garfinkel Presented by Jevin Sweval Introduction to carving File fragmentation Object validation Carving methods Conclusion 1 Carving is the recovery of files from a raw dump of a storage device
More informationVirtual File System -Uniform interface for the OS to see different file systems.
Virtual File System -Uniform interface for the OS to see different file systems. Temporary File Systems -Disks built in volatile storage NFS -file system addressed over network File Allocation -Contiguous
More informationIn-core compression: how to shrink your database size in several times. Aleksander Alekseev Anastasia Lubennikova.
In-core compression: how to shrink your database size in several times Aleksander Alekseev Anastasia Lubennikova www.postgrespro.ru What does Postgres store? Agenda A couple of words about storage internals
More informationCS370: Operating Systems [Spring 2017] Dept. Of Computer Science, Colorado State University
Frequently asked questions from the previous class survey CS 370: OPERATING SYSTEMS [MASS STORAGE] How does the OS caching optimize disk performance? How does file compression work? Does the disk change
More informationAnnouncements. Persistence: Log-Structured FS (LFS)
Announcements P4 graded: In Learn@UW; email 537-help@cs if problems P5: Available - File systems Can work on both parts with project partner Watch videos; discussion section Part a : file system checker
More informationStructuring PLFS for Extensibility
Structuring PLFS for Extensibility Chuck Cranor, Milo Polte, Garth Gibson PARALLEL DATA LABORATORY Carnegie Mellon University What is PLFS? Parallel Log Structured File System Interposed filesystem b/w
More informationVirtual Memory. Kevin Webb Swarthmore College March 8, 2018
irtual Memory Kevin Webb Swarthmore College March 8, 2018 Today s Goals Describe the mechanisms behind address translation. Analyze the performance of address translation alternatives. Explore page replacement
More informationSDS: A Framework for Scientific Data Services
SDS: A Framework for Scientific Data Services Bin Dong, Suren Byna*, John Wu Scientific Data Management Group Lawrence Berkeley National Laboratory Finding Newspaper Articles of Interest Finding news articles
More informationWhat is a file system
COSC 6397 Big Data Analytics Distributed File Systems Edgar Gabriel Spring 2017 What is a file system A clearly defined method that the OS uses to store, catalog and retrieve files Manage the bits that
More informationFROM 4D WRITE TO 4D WRITE PRO INTRODUCTION. Presented by: Achim W. Peschke
4 D S U M M I T 2 0 1 8 FROM 4D WRITE TO 4D WRITE PRO Presented by: Achim W. Peschke INTRODUCTION In this session we will talk to you about the new 4D Write Pro. I think in between everyone knows what
More informationCS370 Operating Systems
CS370 Operating Systems Colorado State University Yashwant K Malaiya Spring 2018 Lecture 24 Mass Storage, HDFS/Hadoop Slides based on Text by Silberschatz, Galvin, Gagne Various sources 1 1 FAQ What 2
More informationCS162 Operating Systems and Systems Programming Lecture 12. Address Translation. Page 1
CS162 Operating Systems and Systems Programming Lecture 12 Translation March 10, 2008 Prof. Anthony D. Joseph http://inst.eecs.berkeley.edu/~cs162 Review: Important Aspects of Memory Multiplexing Controlled
More informationDistributed Systems. Lec 10: Distributed File Systems GFS. Slide acks: Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung
Distributed Systems Lec 10: Distributed File Systems GFS Slide acks: Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung 1 Distributed File Systems NFS AFS GFS Some themes in these classes: Workload-oriented
More informationI/O: State of the art and Future developments
I/O: State of the art and Future developments Giorgio Amati SCAI Dept. Rome, 18/19 May 2016 Some questions Just to know each other: Why are you here? Which is the typical I/O size you work with? GB? TB?
More informationDataMods. Programmable File System Services. Noah Watkins
DataMods Programmable File System Services Noah Watkins jayhawk@cs.ucsc.edu 1 Overview Middleware scalability challenges How DataMods can help address some issues Case Studies Hadoop Cloud I/O Stack Offline
More informationGFS Overview. Design goals/priorities Design for big-data workloads Huge files, mostly appends, concurrency, huge bandwidth Design for failures
GFS Overview Design goals/priorities Design for big-data workloads Huge files, mostly appends, concurrency, huge bandwidth Design for failures Interface: non-posix New op: record appends (atomicity matters,
More informationSelection Queries. to answer a selection query (ssn=10) needs to traverse a full path.
Hashing B+-tree is perfect, but... Selection Queries to answer a selection query (ssn=) needs to traverse a full path. In practice, 3-4 block accesses (depending on the height of the tree, buffering) Any
More informationOperating Systems 2010/2011
Operating Systems 2010/2011 Input/Output Systems part 2 (ch13, ch12) Shudong Chen 1 Recap Discuss the principles of I/O hardware and its complexity Explore the structure of an operating system s I/O subsystem
More informationCost Models for Query Processing Strategies in the Active Data Repository
Cost Models for Query rocessing Strategies in the Active Data Repository Chialin Chang Institute for Advanced Computer Studies and Department of Computer Science University of Maryland, College ark 272
More information朱义普. Resolving High Performance Computing and Big Data Application Bottlenecks with Application-Defined Flash Acceleration. Director, North Asia, HPC
October 28, 2013 Resolving High Performance Computing and Big Data Application Bottlenecks with Application-Defined Flash Acceleration 朱义普 Director, North Asia, HPC DDN Storage Vendor for HPC & Big Data
More informationLustre on ZFS. At The University of Wisconsin Space Science and Engineering Center. Scott Nolin September 17, 2013
Lustre on ZFS At The University of Wisconsin Space Science and Engineering Center Scott Nolin September 17, 2013 Why use ZFS for Lustre? The University of Wisconsin Space Science and Engineering Center
More informationExample Implementations of File Systems
Example Implementations of File Systems Last modified: 22.05.2017 1 Linux file systems ext2, ext3, ext4, proc, swap LVM Contents ZFS/OpenZFS NTFS - the main MS Windows file system 2 Linux File Systems
More informationDistributed File Systems II
Distributed File Systems II To do q Very-large scale: Google FS, Hadoop FS, BigTable q Next time: Naming things GFS A radically new environment NFS, etc. Independence Small Scale Variety of workloads Cooperation
More informationUsing file systems at HC3
Using file systems at HC3 Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association www.kit.edu Basic Lustre
More informationCS435 Introduction to Big Data FALL 2018 Colorado State University. 11/7/2018 Week 12-B Sangmi Lee Pallickara. FAQs
11/7/2018 CS435 Introduction to Big Data - FALL 2018 W12.B.0.0 CS435 Introduction to Big Data 11/7/2018 CS435 Introduction to Big Data - FALL 2018 W12.B.1 FAQs Deadline of the Programming Assignment 3
More informationHPC Input/Output. I/O and Darshan. Cristian Simarro User Support Section
HPC Input/Output I/O and Darshan Cristian Simarro Cristian.Simarro@ecmwf.int User Support Section Index Lustre summary HPC I/O Different I/O methods Darshan Introduction Goals Considerations How to use
More informationCOSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters
COSC 6374 Parallel I/O (I) I/O basics Fall 2010 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network card 1 Network card
More informationpnfs support for ONTAP Unstriped file systems (WIP) Pranoop Erasani Connectathon Feb 22, 2010
pnfs support for ONTAP Unstriped file systems (WIP) Pranoop Erasani pranoop@netapp.com Connectathon Feb 22, 2010 Agenda Clustered ONTAP Architecture Striped WAFL pnfs and Striped WAFL Unstriped WAFL pnfs
More informationCS 318 Principles of Operating Systems
CS 318 Principles of Operating Systems Fall 2017 Lecture 16: File Systems Examples Ryan Huang File Systems Examples BSD Fast File System (FFS) - What were the problems with the original Unix FS? - How
More informationToday s Papers. Array Reliability. RAID Basics (Two optional papers) EECS 262a Advanced Topics in Computer Systems Lecture 3
EECS 262a Advanced Topics in Computer Systems Lecture 3 Filesystems (Con t) September 10 th, 2012 John Kubiatowicz and Anthony D. Joseph Electrical Engineering and Computer Sciences University of California,
More informationUnique Data Organization
Unique Data Organization INTRODUCTION Apache CarbonData stores data in the columnar format, with each data block sorted independently with respect to each other to allow faster filtering and better compression.
More informationCSCI-GA Operating Systems. I/O : Disk Scheduling and RAID. Hubertus Franke
CSCI-GA.2250-001 Operating Systems I/O : Disk Scheduling and RAID Hubertus Franke frankeh@cs.nyu.edu Disks Scheduling Abstracted by OS as files A Conventional Hard Disk (Magnetic) Structure Hard Disk
More informationIME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning
IME (Infinite Memory Engine) Extreme Application Acceleration & Highly Efficient I/O Provisioning September 22 nd 2015 Tommaso Cecchi 2 What is IME? This breakthrough, software defined storage application
More informationREDCap Importing and Exporting (302)
REDCap Importing and Exporting (302) Learning objectives Report building Exporting data from REDCap Importing data into REDCap Backup options API Basics ITHS Focus Speeding science to clinical practice
More informationMemory Management. Disclaimer: some slides are adopted from book authors slides with permission 1
Memory Management Disclaimer: some slides are adopted from book authors slides with permission 1 CPU management Roadmap Process, thread, synchronization, scheduling Memory management Virtual memory Disk
More informationChapter 12: File System Implementation
Chapter 12: File System Implementation Virtual File Systems. Allocation Methods. Folder Implementation. Free-Space Management. Directory Block Placement. Recovery. Virtual File Systems An object-oriented
More informationMemory Management. Today. Next Time. Basic memory management Swapping Kernel memory allocation. Virtual memory
Memory Management Today Basic memory management Swapping Kernel memory allocation Next Time Virtual memory Midterm results Average 68.9705882 Median 70.5 Std dev 13.9576965 12 10 8 6 4 2 0 [0,10) [10,20)
More informationI/O in scientific applications
COSC 4397 Parallel I/O (II) Access patterns Spring 2010 I/O in scientific applications Different classes of I/O operations Required I/O: reading input data and writing final results Checkpointing: data
More informationLustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG
Lustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG 2017 @Beijing Outline LNet reliability DNE improvements Small file performance File Level Redundancy Miscellaneous improvements
More informationFrom server-side to host-side:
From server-side to host-side: Flash memory for enterprise storage Jiri Schindler et al. (see credits) Advanced Technology Group NetApp May 9, 2012 v 1.0 Data Centers with Flash SSDs iscsi/nfs/cifs Shared
More informationParallel I/O Performance Study and Optimizations with HDF5, A Scientific Data Package
Parallel I/O Performance Study and Optimizations with HDF5, A Scientific Data Package MuQun Yang, Christian Chilan, Albert Cheng, Quincey Koziol, Mike Folk, Leon Arber The HDF Group Champaign, IL 61820
More informationFastScale: Accelerate RAID Scaling by
FastScale: Accelerate RAID Scaling by Minimizing i i i Data Migration Weimin Zheng, Guangyan Zhang gyzh@tsinghua.edu.cn Tsinghua University Outline Motivation Minimizing data migration Optimizing data
More informationNDF data format and API implementation. Dr. Bojian Liang Department of Computer Science University of York 25 March 2010
NDF data format and API implementation Dr. Bojian Liang Department of Computer Science University of York 25 March 2010 What is NDF? Slide 2 The Neurophysiology Data translation Format (NDF) is a data
More informationData Management. Parallel Filesystems. Dr David Henty HPC Training and Support
Data Management Dr David Henty HPC Training and Support d.henty@epcc.ed.ac.uk +44 131 650 5960 Overview Lecture will cover Why is IO difficult Why is parallel IO even worse Lustre GPFS Performance on ARCHER
More informationZettabyte Reliability with Flexible End-to-end Data Integrity
Zettabyte Reliability with Flexible End-to-end Data Integrity Yupu Zhang, Daniel Myers, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau University of Wisconsin - Madison 5/9/2013 1 Data Corruption Imperfect
More informationLecture 16: Storage Devices
CS 422/522 Design & Implementation of Operating Systems Lecture 16: Storage Devices Zhong Shao Dept. of Computer Science Yale University Acknowledgement: some slides are taken from previous versions of
More informationChapter 12: File System Implementation
Chapter 12: File System Implementation Silberschatz, Galvin and Gagne 2013 Chapter 12: File System Implementation File-System Structure File-System Implementation Allocation Methods Free-Space Management
More informationChapter 10: Mass-Storage Systems
Chapter 10: Mass-Storage Systems Silberschatz, Galvin and Gagne 2013 Chapter 10: Mass-Storage Systems Overview of Mass Storage Structure Disk Structure Disk Attachment Disk Scheduling Disk Management Swap-Space
More informationOS and Hardware Tuning
OS and Hardware Tuning Tuning Considerations OS Threads Thread Switching Priorities Virtual Memory DB buffer size File System Disk layout and access Hardware Storage subsystem Configuring the disk array
More informationLustre and PLFS Parallel I/O Performance on a Cray XE6
Lustre and PLFS Parallel I/O Performance on a Cray XE6 Cray User Group 2014 Lugano, Switzerland May 4-8, 2014 April 2014 1 Many currently contributing to PLFS LANL: David Bonnie, Aaron Caldwell, Gary Grider,
More informationThe HDF Group. Parallel HDF5. Extreme Scale Computing Argonne.
The HDF Group Parallel HDF5 Advantage of Parallel HDF5 Recent success story Trillion particle simulation on hopper @ NERSC 120,000 cores 30TB file 23GB/sec average speed with 35GB/sec peaks (out of 40GB/sec
More informationChapter 10: Mass-Storage Systems. Operating System Concepts 9 th Edition
Chapter 10: Mass-Storage Systems Silberschatz, Galvin and Gagne 2013 Chapter 10: Mass-Storage Systems Overview of Mass Storage Structure Disk Structure Disk Attachment Disk Scheduling Disk Management Swap-Space
More informationORC Files. Owen O June Page 1. Hortonworks Inc. 2012
ORC Files Owen O Malley owen@hortonworks.com @owen_omalley owen@hortonworks.com June 2013 Page 1 Who Am I? First committer added to Hadoop in 2006 First VP of Hadoop at Apache Was architect of MapReduce
More information<Insert Picture Here> Linux Data Integrity
Linux Data Integrity Martin K. Petersen Consulting Software Developer, Linux Engineering Topics DIF/DIX Data Corruption T10 DIF Data Integrity Extensions Linux & Data Integrity Block
More information