Agenda. Project #3, Part I. Agenda. Project #3, Part I. No SSE 12/1/10

Similar documents
Agenda. Agenda 11/12/12. Review - 6 Great Ideas in Computer Architecture

CS2410: Computer Architecture. Storage systems. Sangyeun Cho. Computer Science Department University of Pittsburgh

Computer Science 146. Computer Architecture

Today: Coda, xfs! Brief overview of other file systems. Distributed File System Requirements!

Lecture 23: I/O Redundant Arrays of Inexpensive Disks Professor Randy H. Katz Computer Science 252 Spring 1996

Lecture 25: Dependability and RAID

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Dependability and RAID. Great Ideas in Computer Architecture

Dependency Redundancy & RAID

CS61C : Machine Structures

CS 61C: Great Ideas in Computer Architecture Lecture 25: Dependability and RAID

Agenda. Recap: Components of a Computer. Agenda. Recap: Cache Performance and Average Memory Access Time (AMAT) Recap: Typical Memory Hierarchy

CS 61C: Great Ideas in Computer Architecture. Dependability. Guest Lecturer: Paul Ruan

Today: Coda, xfs. Case Study: Coda File System. Brief overview of other file systems. xfs Log structured file systems HDFS Object Storage Systems

Advanced Computer Architecture

CSE325 Principles of Operating Systems. Mass-Storage Systems. David P. Duggan. April 19, 2011

CS61C : Machine Structures

CS61C - Machine Structures. Week 7 - Disks. October 10, 2003 John Wawrzynek

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Computer Architecture ECE 568

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Computer Architecture ECE 568

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Computer Architecture ECE 568

CS61C : Machine Structures

UCB CS61C : Machine Structures

NOW Handout Page 1 S258 S99 1

www-inst.eecs.berkeley.edu/~cs61c/

Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted from Hennessy & Patterson / 2003 Elsevier

Definition of RAID Levels

HP AutoRAID (Lecture 5, cs262a)

I/O CANNOT BE IGNORED

Reliable Computing I

Administrivia. CMSC 411 Computer Systems Architecture Lecture 19 Storage Systems, cont. Disks (cont.) Disks - review

Appendix D: Storage Systems

Page 1. Motivation: Who Cares About I/O? Lecture 5: I/O Introduction: Storage Devices & RAID. I/O Systems. Big Picture: Who cares about CPUs?

Topics. Lecture 8: Magnetic Disks

Chapter 6. I/O issues

Review. Lecture #25 Input / Output, Networks II, Disks ABCs of Networks: 2 Computers

Lecture: Storage, GPUs. Topics: disks, RAID, reliability, GPUs (Appendix D, Ch 4)

I/O CANNOT BE IGNORED

HP AutoRAID (Lecture 5, cs262a)

Recap of Networking Intro. Lecture #28 Networking & Disks Protocol Family Concept. Protocol for Networks of Networks?

RAID (Redundant Array of Inexpensive Disks)

Concepts Introduced. I/O Cannot Be Ignored. Typical Collection of I/O Devices. I/O Issues

Today s Papers. Array Reliability. RAID Basics (Two optional papers) EECS 262a Advanced Topics in Computer Systems Lecture 3

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Mul$- Level Caches

Storage Systems. Storage Systems

Page 1. Magnetic Disk Purpose Long term, nonvolatile storage Lowest level in the memory hierarchy. Typical Disk Access Time

CS5460: Operating Systems Lecture 20: File System Reliability

Lecture for Storage System

Storage systems. Computer Systems Architecture CMSC 411 Unit 6 Storage Systems. (Hard) Disks. Disk and Tape Technologies. Disks (cont.

Address Accessible Memories. A.R. Hurson Department of Computer Science Missouri University of Science & Technology

Lecture 21: Reliable, High Performance Storage. CSC 469H1F Fall 2006 Angela Demke Brown

Lecture 23: Storage Systems. Topics: disk access, bus design, evaluation metrics, RAID (Sections )

Chapter 6. Storage and Other I/O Topics

- SLED: single large expensive disk - RAID: redundant array of (independent, inexpensive) disks

CSE 502 Graduate Computer Architecture. Lec 22 Disk Storage

CS370: System Architecture & Software [Fall 2014] Dept. Of Computer Science, Colorado State University

RAID. Redundant Array of Inexpensive Disks. Industry tends to use Independent Disks

5.11 Parallelism and Memory Hierarchy: Redundant Arrays of Inexpensive Disks 485.e1

Virtual Memory. Reading. Sections 5.4, 5.5, 5.6, 5.8, 5.10 (2) Lecture notes from MKP and S. Yalamanchili

Outline. EEL-4713 Computer Architecture I/O Systems. I/O System Design Issues. The Big Picture: Where are We Now? I/O Performance Measures

Mladen Stefanov F48235 R.A.I.D

ECE Enterprise Storage Architecture. Fall 2018

CS 341l Fall 2008 Test #4 NAME: Key

CMSC 424 Database design Lecture 12 Storage. Mihai Pop

CS61C : Machine Structures

CSE 153 Design of Operating Systems

In the late 1980s, rapid adoption of computers

IST346. Data Storage

RAID SEMINAR REPORT /09/2004 Asha.P.M NO: 612 S7 ECE

Database Systems. November 2, 2011 Lecture #7. topobo (mit)

Computer Architecture 计算机体系结构. Lecture 6. Data Storage and I/O 第六讲 数据存储和输入输出. Chao Li, PhD. 李超博士

Chapter 12: Mass-Storage

CS 61C: Great Ideas in Computer Architecture Case Studies: Server and Cellphone microprocessors. Not yet in producvon, the next core awer Ivy Bridge!

CSC Operating Systems Spring Lecture - XIX Storage and I/O - II. Tevfik Koşar. Louisiana State University.

RAID Structure. RAID Levels. RAID (cont) RAID (0 + 1) and (1 + 0) Tevfik Koşar. Hierarchical Storage Management (HSM)

Today: Secondary Storage! Typical Disk Parameters!

RAID (Redundant Array of Inexpensive Disks) I/O Benchmarks ABCs of UNIX File Systems A Study Comparing UNIX File System Performance

Principles of Data Management. Lecture #2 (Storing Data: Disks and Files)

CS152 Computer Architecture and Engineering Lecture 25. I/O and Storage Systems Power

Database Systems II. Secondary Storage

Advanced Database Systems

Database Management Systems, 2nd edition, Raghu Ramakrishnan, Johannes Gehrke, McGraw-Hill

CSE 451: Operating Systems Spring Module 18 Redundant Arrays of Inexpensive Disks (RAID)

Storage. Hwansoo Han

CSE 120. Operating Systems. March 27, 2014 Lecture 17. Mass Storage. Instructor: Neil Rhodes. Wednesday, March 26, 14

High Performance Computing Course Notes High Performance Storage

Instructors: Randy H. Katz David A. PaGerson hgp://inst.eecs.berkeley.edu/~cs61c/fa10. 10/4/10 Fall Lecture #16. Agenda

CSCI-GA Operating Systems. I/O : Disk Scheduling and RAID. Hubertus Franke

CS 61C: Great Ideas in Computer Architecture. Direct Mapped Caches

CSE 124: Networked Services Lecture-17

UC Berkeley CS61C : Machine Structures

CS3600 SYSTEMS AND NETWORKS

Che-Wei Chang Department of Computer Science and Information Engineering, Chang Gung University

CS 351 Exam 3 Mon. 5/11/2015

Storage. CS 3410 Computer System Organization & Programming

10/19/17. You Are Here! Review: Direct-Mapped Cache. Typical Memory Hierarchy

Review: Performance Latency vs. Throughput. Time (seconds/program) is performance measure Instructions Clock cycles Seconds.

Frequently asked questions from the previous class survey

CSE 451: Operating Systems Winter Redundant Arrays of Inexpensive Disks (RAID) and OS structure. Gary Kimura

Guerrilla 7: Virtual Memory, ECC, IO

SYSTEM UPGRADE, INC Making Good Computers Better. System Upgrade Teaches RAID

Transcription:

// CS 6C: Great Ideas in Computer Architecture (Machine Structures) Project Speed- up and RAID Instructors: Randy H. Katz David A. PaGerson hgp://inst.eecs.berkeley.edu/~cs6c/fa // Fall - - Lecture #9 // Fall - - Lecture #9 Project #, Part I sgemm (x) levels blocking, loop unrolling Class Data Peak Simple Blocked Thanks TA Andrew! // Fall - - Lecture #9 // Fall - - Lecture # Project #, Part I No SSE Num. Submissions sgemm (x) Performance Histogram sgemm (x) levels blocking, loop unrolling Thanks TA Andrew! // Fall - - Lecture # // Fall - - Lecture #9 6

// No SSE No SSE sgemm (x, Peak: 7) levels blocking, loop unrolling sgemm (x) Performance Histogram.9.8 % peak.7.6..... Num. Submissions // Fall - - Lecture #9 7 // Fall - - Lecture #9 8 8 6 SSE sgemm SSE ( thread) A = x A'A.8.7 SSE % of Goto Single-thread x A'A 8 6 Simple Blocked GotoBLAS CS6c % Goto (7. ).6..... // Fall - - Lecture #9 9 // Fall - - Lecture #9 SSE OMP Histogram of Performance (SSE) A = x sgemm OMP (8 threads,a'a) A = x # submissions 8 6 8 6 Naive GotoBLAS CS6c...... 6. 7. 8. 9..... // Fall - - Lecture #9 // Fall - - Lecture #9

// OMP OMP.9 % of Goto OMP (8 threads) x Histogram of performance (OMP) A = x % Goto ( ).8.7.6... # submissions 9 8 7 6.. 6 7 8 9 // Fall - - Lecture #9 // Fall - - Lecture #9 EC and Part (serial) A'A (x) EC and Part (8 threads) A'A (x) EC Part EC Part // Fall - - Lecture #9 // Fall - - Lecture #9 6 Performance Change (serial) A'A (x) Performance Change (8 threads) A'A (x). Speedup (EC/Part).. Speedup (EC/Part)... // Fall - - Lecture #9 7 // Fall - - Lecture #9 8

// Administrivia Final Review: Mon Dec 6, - PM, Evans Final: Mon Dec 8- AM ( Hearst Gym) Like midterm: T/F, M/C, short answers Whole Course: readings, lectures, projects, labs, hw Emphasize nd half of 6C + midterm mistakes // Fall - - Lecture #9 9 // Fall - - Lecture #9 Cloud Compucng Datacenter Economics // Fall - - Lecture #9 // Fall - - Lecture #9 Evolucon of the Disk Drive Arrays of Small Disks Can smaller disks be used to close gap in performance between disks and CPUs? IBM 9K, 986 Convenconal: disk designs.. Low End High End IBM RAMAC, 96 Apple SCSI, 986 // Fall - - Lecture #9 Disk Array: disk design. // Fall - - Lecture #9

// Replace Small Number of Large Disks with Large Number of Small Disks! (988 Disks) IBM 9K IBM." 6 x7 GBytes MBytes GBytes Capacity Volume Power Data Rate I/O Rate MTTF Cost 97 cu. i. KW MB/s 6 I/Os/s KHrs $K. cu. i. W. MB/s I/Os/s KHrs $K cu. i. KW MB/s 9 IOs/s??? Hrs $K Disk Arrays have potencal for large data and I/O rates, high MB per cu. i., high MB per KW, but what about reliability? // Fall - - Lecture #9 9X X 8X 6X RAID: Redundant Arrays of (Inexpensive) Disks Files are "striped" across mulcple disks Redundancy yields high data availability Availability: service scll provided to user, even if some components failed Disks will scll fail Contents reconstructed from data redundantly stored in the array Capacity penalty to store redundant info Bandwidth penalty to update redundant info // Fall - - Lecture #9 6 Redundant Arrays of Inexpensive Disks RAID : Disk Mirroring/Shadowing recovery group Each disk is fully duplicated onto its mirror Very high availability can be achieved Bandwidth sacrifice on write: Logical write = two physical writes Reads may be opcmized Most expensive solucon: % capacity overhead // Fall - - Lecture #9 7 Redundant Array of Inexpensive Disks RAID : Parity Disk... logical record Striped physical records P contains sum of other disks per stripe mod ( parity ) If disk fails, subtract P from sum of other disks to find missing informacon // Fall - - Lecture #9 8 P Redundant Arrays of Inexpensive Disks RAID : High I/O Rate Parity Insides of disks Example: small read D & D, large write D- D D D D D P D D D6 D7 P D8 D9 D D P D D D D D6 D7 D8 D9 D D D D P."."."."."." Disk." Columns."."." //." Fall." - - Lecture." #9."." 9 P P Increasing Logical Disk Address Stripe Inspiracon for RAID works well for small reads Small writes (write to one disk): Opcon : read other data disks, create new sum and write to Parity Disk Opcon : since P has old sum, compare old data to new data, add the difference to P Small writes are limited by Parity Disk: Write to D, D both also write to P disk D D D D P D D D6 D7 P // Fall - - Lecture #9

// RAID : High I/O Rate Interleaved Parity Problems of Disk Arrays: Small Writes RAID- : Small Write Algorithm D Independent writes possible because of interleaved parity D D D D D D6 D7 D8 D9 D D Increasing Logical Disk Addresses Logical Write = Physical Reads + Physical Writes D' D D D D old data (Read) new data old (Read) parity + XOR D D D D D6 D7 D8 D9 D D D D + XOR Example: write to D, D uses disks,,, // (Write) Disk Columns - - Lecture #9 Fall D' // Tech Report Read Round the World (Write) D D D Fall - - Lecture #9 P' RAID- I (December 987) - I (989) Consisted of a Sun /8 workstacon with 8 MB of DRAM, four dual- string SCSI controllers, 8.- inch SCSI disks and specialized disk striping soiware // Fall - - Lecture #9 // Fall - - Lecture #9 RAID II RAID II 99-99 Early Network AGached Storage (NAS) System running a Log Structured File System (LFS) Impact: $ Billion/year in Over $ Billion in RAID device sold since 99- + RAID companies (at the peak) Soiware RAID a standard component of modern OSs // Fall - - Lecture #9 // Fall - - Lecture #9 6 6

// Summary Logical- to- physical block mapping, parity striping, read- modify- write processing Embedded caches and orchestracng data staging between network interfaces, parity hardware, and file server interfaces Failed disk replacement, hot spares, background copies and backup Embedded log- structured file systems, compression on the fly Soiware complexity dominates hardware! // Fall - - Lecture #9 7 7