CS61C - Machine Structures. Week 7 - Disks. October 10, 2003 John Wawrzynek

Similar documents
www-inst.eecs.berkeley.edu/~cs61c/

CS61C : Machine Structures

UCB CS61C : Machine Structures

CS2410: Computer Architecture. Storage systems. Sangyeun Cho. Computer Science Department University of Pittsburgh

Topics. Lecture 8: Magnetic Disks

Page 1. Motivation: Who Cares About I/O? Lecture 5: I/O Introduction: Storage Devices & RAID. I/O Systems. Big Picture: Who cares about CPUs?

High-Performance Storage Systems

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Computer Architecture ECE 568

CS61C : Machine Structures

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Computer Architecture ECE 568

Recap of Networking Intro. Lecture #28 Networking & Disks Protocol Family Concept. Protocol for Networks of Networks?

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Computer Architecture ECE 568

Review. Lecture #25 Input / Output, Networks II, Disks ABCs of Networks: 2 Computers

Computer Science 61C Spring Friedland and Weaver. Input/Output

Computer Science 146. Computer Architecture

CHAPTER 12: MASS-STORAGE SYSTEMS (A) By I-Chen Lin Textbook: Operating System Concepts 9th Ed.

Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted from Hennessy & Patterson / 2003 Elsevier

CS61C : Machine Structures

Computer Science 146. Computer Architecture

Most important topics today. I/O bus (computer bus) Polling vs interrupts Switch vs shared network Network protocol layers Various types of RAID

Lecture #28 I/O, networks, disks

Lecture for Storage System

CS152 Computer Architecture and Engineering Lecture 19: I/O Systems

I/O CANNOT BE IGNORED

Advanced Computer Architecture

I/O CANNOT BE IGNORED

Chapter 6. Storage and Other I/O Topics

Lecture 23: I/O Redundant Arrays of Inexpensive Disks Professor Randy H. Katz Computer Science 252 Spring 1996

CS61C - Machine Structures. Lecture 17 - Caches, Part I. October 25, 2000 David Patterson

Computer System Architecture

SCSI vs. ATA More than an interface. Dave Anderson, Jim Dykes, Erik Riedel Seagate Research April 2003

NOW Handout Page 1 S258 S99 1

inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures Lecture 26 IO Basics, Storage & Performance I ! HACKING THE SMART GRID!

Goals for today:! Recall : 5 components of any Computer! ! What do we need to make I/O work?!

CPS104 Computer Organization and Programming Lecture 18: Input-Output. Outline of Today s Lecture. The Big Picture: Where are We Now?

Concepts Introduced. I/O Cannot Be Ignored. Typical Collection of I/O Devices. I/O Issues

UCB CS61C : Machine Structures

Motivation for Input/Output

Agenda. Project #3, Part I. Agenda. Project #3, Part I. No SSE 12/1/10

Appendix D: Storage Systems

Storage Systems. Storage Systems

ECE232: Hardware Organization and Design

Today: Coda, xfs! Brief overview of other file systems. Distributed File System Requirements!

Storage System COSC UCB

Components of the Virtual Memory System

CS 261 Fall Mike Lam, Professor. Memory

CMSC 611: Advanced Computer Architecture

Where Have We Been? Ch. 6 Memory Technology

Virtual Memory. Reading. Sections 5.4, 5.5, 5.6, 5.8, 5.10 (2) Lecture notes from MKP and S. Yalamanchili

CSE 153 Design of Operating Systems

Chapter 6. Storage & Other I/O

COS 318: Operating Systems. Storage Devices. Vivek Pai Computer Science Department Princeton University

Agenda. Agenda 11/12/12. Review - 6 Great Ideas in Computer Architecture

CSE 451: Operating Systems Spring Module 12 Secondary Storage

Page 1. Magnetic Disk Purpose Long term, nonvolatile storage Lowest level in the memory hierarchy. Typical Disk Access Time

Session: Hardware Topic: Disks. Daniel Chang. COP 3502 Introduction to Computer Science. Lecture. Copyright August 2004, Daniel Chang

CS252 S05. CMSC 411 Computer Systems Architecture Lecture 18 Storage Systems 2. I/O performance measures. I/O performance measures

A track on a magnetic disk is a concentric rings where data is stored.

COS 318: Operating Systems. Storage Devices. Jaswinder Pal Singh Computer Science Department Princeton University

Storage Technologies and the Memory Hierarchy

CSCI-UA.0201 Computer Systems Organization Memory Hierarchy

EXTERNAL MEMORY (Part 1)

Storage systems. Computer Systems Architecture CMSC 411 Unit 6 Storage Systems. (Hard) Disks. Disk and Tape Technologies. Disks (cont.

CS 201 The Memory Hierarchy. Gerson Robboy Portland State University

Wednesday, April 25, Discs RAID: Introduction Error detection and correction Error detection: Simple parity Error correction: Hamming Codes

CSE 153 Design of Operating Systems Fall 2018

Contents. Memory System Overview Cache Memory. Internal Memory. Virtual Memory. Memory Hierarchy. Registers In CPU Internal or Main memory

Show how to connect three Full Adders to implement a 3-bit ripple-carry adder

COS 318: Operating Systems. Storage Devices. Kai Li Computer Science Department Princeton University

CSE 451: Operating Systems Spring Module 12 Secondary Storage. Steve Gribble

Quiz for Chapter 6 Storage and Other I/O Topics 3.10

Monday, May 4, Discs RAID: Introduction Error detection and correction Error detection: Simple parity Error correction: Hamming Codes

CS 33. Memory Hierarchy I. CS33 Intro to Computer Systems XVI 1 Copyright 2016 Thomas W. Doeppner. All rights reserved.

Outline. EEL-4713 Computer Architecture I/O Systems. I/O System Design Issues. The Big Picture: Where are We Now? I/O Performance Measures

Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted from Hennessy & Patterson / 2003 Elsevier

Introduction to Input and Output

CS 471 Operating Systems. Yue Cheng. George Mason University Fall 2017

Storage. CS 3410 Computer System Organization & Programming

Reading and References. Input / Output. Why Input and Output? A typical organization. CSE 410, Spring 2004 Computer Systems

Administrivia. CMSC 411 Computer Systems Architecture Lecture 19 Storage Systems, cont. Disks (cont.) Disks - review

Readings. Storage Hierarchy III: I/O System. I/O (Disk) Performance. I/O Device Characteristics. often boring, but still quite important

Database Systems II. Secondary Storage

CS61C : Machine Structures

COMP283-Lecture 3 Applied Database Management

Storage Devices for Database Systems

Hard Disk Drives. Nima Honarmand (Based on slides by Prof. Andrea Arpaci-Dusseau)

ECE331: Hardware Organization and Design

Key Points. Rotational delay vs seek delay Disks are slow. Techniques for making disks faster. Flash and SSDs

Module 1: Basics and Background Lecture 4: Memory and Disk Accesses. The Lecture Contains: Memory organisation. Memory hierarchy. Disks.

Semiconductor Memory Types Microprocessor Design & Organisation HCA2102

CENG3420 Lecture 08: Memory Organization

UNIT 2 Data Center Environment

CS 261 Fall Mike Lam, Professor. Memory

The Fetch-Execute Cycle

Chapter 6 External Memory

William Stallings Computer Organization and Architecture 8 th Edition. Chapter 6 External Memory

The Memory Hierarchy /18-213/15-513: Introduction to Computer Systems 11 th Lecture, October 3, Today s Instructor: Phil Gibbons

CS162 Operating Systems and Systems Programming Lecture 17. Disk Management and File Systems

CSE 451: Operating Systems Winter Secondary Storage. Steve Gribble. Secondary storage

u Covered: l Management of CPU & concurrency l Management of main memory & virtual memory u Currently --- Management of I/O devices

Transcription:

CS61C - Machine Structures Week 7 - Disks October 10, 2003 John Wawrzynek http://www-inst.eecs.berkeley.edu/~cs61c/ CS61C wk 7 Disks UC Regents 1

Magnetic Disk common I/O device Computer Processor (active) Control ( brain ) Datapath ( brawn ) Memory (passive) (where programs, data live when running) Devices Input Output Keyboard, Mouse Disk, Network Display, Printer CS61C wk 7 Disks UC Regents 2

Magnetic Disk common I/O device A kind of Computer Memory: Information sorted by magnetizing ferrite material on surface of rotating disk (similar to tape recorder except binary signaling). Nonvolatile storage retains its value without applying power to disk. Types: Floppy disks slower, less dense, removable. Hard Disk Drives faster, more dense, non-removable. Purpose in computer systems (Hard Drive): 1. Long-term, inexpensive storage for files 2. Backup for main-memory. Large, inexpensive, slow level in the memory hierarchy (discuss later) CS61C wk 7 Disks UC Regents 3

CS61C wk 7 Disks UC Regents 4

Hard Drives are Sealed. Why? The closer the head to the disk, the small the spot size and thus the denser the recording. Measured in Gbit/in 2. ~60 is state of the art. Disks are sealed to keep the dust out. Heads are designed to fly at around 5-20nm above the surface of the disk. 99.999% of the head/arm weight is supported by the air bearing force (air cushion) developed between the disk and the head. CS61C wk 7 Disks UC Regents 5

Photo of Disk Head, Arm, Actuator Arm Spindle Actuator Head Platters (12) CS61C wk 7 Disks UC Regents 6

Disk Device Terminology Arm Head Sector Inner Track Outer Track Actuator Platter Several platters, with information recorded magnetically on both surfaces (usually) Bits recorded in tracks, which in turn divided into sectors (e.g., 512 Bytes) Actuator moves head (end of arm) over track ( seek ), wait for sector rotate under head, then read or write CS61C wk 7 Disks UC Regents 7

Disk Device Performance Platter Outer Track InnerSector Head Arm Track Spindle Controller Actuator Disk Latency = Seek Time + Rotation Time + Transfer Time + Controller Overhead Seek Time? depends no. tracks move arm, seek speed of disk Rotation Time? depends on speed disk rotates, how far sector is from head Transfer Time? depends on data rate (bandwidth) of disk (f(bit density,rpm)), size of request CS61C wk 7 Disks UC Regents 8

Disk Device Performance Average distance of sector from head? 1/2 time of a rotation 7200 Revolutions Per Minute 120 Rev/sec 1 revolution = 1/120 sec 8.33 milliseconds 1/2 rotation (revolution) 4.17 ms Average no. tracks to move arm? Disk industry standard benchmark: Sum all time for all possible seek distances from all possible tracks / # possible Assumes average seek distance is random CS61C wk 7 Disks UC Regents 9

Data Rate: Inner vs. Outer Tracks To keep things simple, originally same number of sectors per track Since outer track longer, lower bits per inch Competition decided to keep bits per inch (BPI) high for all tracks ( constant bit density ) More capacity per disk More sectors per track towards edge Since disk spins at constant speed, outer tracks have faster data rate Bandwidth outer track 1.7X inner track! CS61C wk 7 Disks UC Regents 10

Disk Performance Model /Trends Capacity + 100%/year (2X / 1.0 yrs) Over time, this has grown so fast that # of platters has reduced (some even use only one now!) Transfer rate (BW) + 40%/year (2X / 2.0 yrs) Rotation + Seek time 8%/year (1/2 in 10 yrs) MB/$ > 100%/year (2X / 1.0 yrs) Fewer chips + bit density improvement CS61C wk 7 Disks UC Regents 11

State of the Art: Ultrastar 73LZX (~2000) source: www.ibm.com; 73.4 GB, 3.5 inch disk 10,000 RPM; 3 ms = 1/2 rotation 6 platters, 12 surfaces 13.2 Gbit/sq. in. areal den 10 watts (idle) 0.1 ms controller time 4.9 ms avg. seek 49 to 87 MB/s(internal) Lower performance disks with similar capacities are available for $3-$9/GB CS61C wk 7 Disks UC Regents 12

State of the Art: Barracuda 180 (~2001) source: www.seagate.com; 181.6 GB, 3.5-inch disk 7200 RPM; SCSI 4.16 ms = 1/2 rotation 12 platters, 24 surfaces 31.2 Gbit/sq. in. areal den 10 watts (idle) 0.1 ms controller time 8.0 ms avg. seek 35 to 64 MB/s(internal) $7.50 / GB (Lower capacity, ATA/IDE disks ~ $2 / GB) CS61C wk 7 Disks UC Regents 13

State of the Art: Deskstar 7K250 (~2002) source: www.hitachi.com; 250 GB, 3.5 inch disk 7200 RPM; 4.17 ms = 1/2 rotation 3 platters, 6 surfaces 62 Gbit/sq. in. areal den 7.0 watts (idle) 0.1 ms controller time 8.5 ms avg. seek Transfer rate: 94.6 MB/s (internal) 56.3 MB/s (external) $166, $1.50/GB Lower capacities disks are available for < $1/GB CS61C wk 7 Disks UC Regents 14

Disk Performance Example Calculate time to read 1 sector (512B) for Deskstar using advertised performance; sector is on outer track Disk latency = average seek time + average rotational delay + transfer time + controller overhead = 8.5 ms + 0.5 * 1/(7200 RPM) + 0.5 KB / (95 MB/s) + 0.1 ms = 8.5 ms + 0.5 /(7200 RPM/(60000ms/M)) + 0.5 KB / (95 KB/ms) + 0.1 ms = 8.5 + 4.17 + 0.0048 + 0.1 ms = 12.77 ms How many CPU clock cycles is this? CS61C wk 7 Disks UC Regents 15

Areal Density Bits recorded along a track Metric is Bits Per Inch (BPI) Number of tracks per surface Metric is Tracks Per Inch (TPI) We care about bit density per unit area Metric is Bits Per Square Inch Called Areal Density Areal Density = BPI x TPI CS61C wk 7 Disks UC Regents 16

Early Disk History (IBM) Data density Mbit/sq. in. Capacity of Unit Shown Megabytes 1973: 1. 7 Mbit/sq. in 140 MBytes 1979: 7. 7 Mbit/sq. in 2,300 MBytes source: New York Times, 2/23/98, page C3, Makers of disk drives crowd even more data into even smaller spaces CS61C wk 7 Disks UC Regents 17

Early Disk History 1989: 63 Mbit/sq. in 60,000 MBytes 1997: 1450 Mbit/sq. in 1600 MBytes 1997: 3090 Mbit/sq. in 8100 MBytes source: New York Times, 2/23/98, page C3, Makers of disk drives crowd even more data into even smaller spaces CS61C wk 7 Disks UC Regents 18

Areal Density Year Areal Density 1973 1.7 1979 7.7 1989 63 1997 3090 2000 17100 2001 31200 2002 62000 100000 10000 Areal Density 1000 100 10 1 1973 1983 1993 2003 Year Areal Density = BPI x TPI Change slope 30%/yr to 60%/yr about 1990 Giant Magnetoresistive effect head (highly sensitive) CS61C wk 7 Disks UC Regents 19

Historical Perspective Form factor and capacity are more important in the marketplace than is performance Form factor evolution: 1970s: Mainframes 14 inch diameter disks 1980s: Minicomputers, Servers 8, 5.25 diameter disks Late 1980s/Early 1990s: PCs 3.5 inch diameter disks Laptops, notebooks 2.5 inch disks Palmtops didn t use disks, so 1.8 inch diameter disks didn t make it CS61C wk 7 Disks UC Regents 20

1 inch disk drive! 2000 IBM MicroDrive: 1.7 x 1.4 x 0.2 1 GB, 3600 RPM, 5 MB/s, 15 ms seek Digital camera, PalmPC? 2006 MicroDrive? 9 GB, 50 MB/s! Assuming it finds a niche in a successful product Assuming past trends continue CS61C wk 7 Disks UC Regents 21

Bonus homework problems Suppose a typical book is 400 pages long, each page on average it has 40 lines of text, and there is an average of line of text has 80 characters. The PDF version of a chapter of our textbook takes 632 KB for 80 pages.the highest capacity 3.5-inch drive today is 250 GB. 1. How many books fit in such a disk today in text? in PDF? How long would it take to access a book from disk? 2. The Kreege Engineering library has 160,000 books and the Bancroft library has 400,000. If disk capacity doubles every year, when would a single disk contain all the books in these two libraries using text? PDF? Might this have impact on the campus? How? CS61C wk 7 Disks UC Regents 22

Bonus slides CS61C wk 7 Disks UC Regents 23

Fallacy: Use Data Sheet Average Seek Time Manufacturers needed standard for fair comparison ( benchmark ) Calculate all seeks from all tracks, divide by number of seeks => average Real average would be based on how data is laid out on disk, and in what order it is accessed by application Usually, tend to seek to tracks nearby, not to random track Rule of Thumb: observed average seek time is typically about 1/4 to 1/3 of quoted seek time (i.e., 3X-4X faster) Deskstar 250 avg. seek: 8.5 ms 2.83 ms CS61C wk 7 Disks UC Regents 24

Fallacy: Use Data Sheet Transfer Rate Manufacturers quote the speed of the data rate off the surface of the disk Sectors contain an error detection and correction field (can be 20% of sector size) plus sector number as well as data There are gaps between sectors on track Rule of Thumb: disks deliver about ~ 1/2-3/4 of internal media rate for data For example, Deskstar 250 quotes 95 MB/s (max) internal media rate v. 56 MB/s (max) external data rate (60%) CS61C wk 7 Disks UC Regents 25

Bonus: Future Disk Size and Performance Continued advance in capacity (60%/yr) and bandwidth (40%/yr) Slow improvement in seek, rotation (8%/yr) Time to read whole disk Year Sequentially Randomly (1 sector/seek) 1990 4 minutes 6 hours 2000 10 minutes 4 days 3.5 form factor make sense in 5-7 yrs? CS61C wk 7 Disks UC Regents 26

Use Arrays of Small Disks? Katz and Patterson asked in 1987: Can smaller disks be used to close gap in performance between disks and CPUs? Conventional: 4 disk designs 3.5 5.25 Low End 10 14 High End Disk Array: 1 disk design 3.5 CS61C wk 7 Disks UC Regents 27

Replace Small Number of Large Disks with Large Number of Small Disks! (1988 Disks) Capacity Volume Power Data Rate I/O Rate MTTF Cost IBM 3390K 20 GBytes 97 cu. ft. 3 KW 15 MB/s 600 I/Os/s 250 KHrs $250K IBM 3.5" 0061 320 MBytes 0.1 cu. ft. 11 W 1.5 MB/s 55 I/Os/s 50 KHrs $2K x70 23 GBytes 11 cu. ft. 1 KW 120 MB/s 3900 IOs/s??? Hrs $150K Disk Arrays potentially high performance, high MB per cu. ft., high MB per KW, but what about reliability? CS61C wk 7 Disks UC Regents 28 9X 3X 8X 6X

Array Reliability Reliability - whether or not a component has failed measured as Mean Time To Failure (MTTF) Reliability of N disks = Reliability of 1 Disk N (assuming failures independent) 50,000 Hours 70 disks = 700 hour Disk system MTTF: Drops from 6 years to 1 month! Disk arrays too unreliable to be useful! CS61C wk 7 Disks UC Regents 29

Redundant Arrays of (Inexpensive) Disks Files are "striped" across multiple disks Redundancy yields high data availability Availability: service still provided to user, even if some components failed When disk fails, contents reconstructed from data redundantly stored in the array Capacity penalty to store redundant info Bandwidth penalty to update redundant info CS61C wk 7 Disks UC Regents 30

Bonus slide: Berkeley History, RAID-I RAID-I (1989) Consisted of a Sun 4/280 workstation with 128 MB of DRAM, four dual-string SCSI controllers, 28 5.25- inch SCSI disks and specialized disk striping software Today RAID is $27 billion dollar industry, 80% nonpc disks sold in RAIDs CS61C wk 7 Disks UC Regents 31

Things to Remember Magnetic Disks continue rapid advance: 60%/yr capacity, 40%/yr bandwidth, slow on seek, rotation improvements, MB/$ improving 100%/yr. Designs to fit high volume form factor Quoted seek times too conservative, data rates too optimistic for use in system RAID Higher performance with more disk arms per $ Adds availability option for small number of extra disks Today RAID is $27 billion dollar industry, 80% nonpc disks sold in RAIDs; started at Cal CS61C wk 7 Disks UC Regents 32