NOW Handout Page 1 S258 S99 1

Similar documents
CS2410: Computer Architecture. Storage systems. Sangyeun Cho. Computer Science Department University of Pittsburgh

Advanced Computer Architecture

Case for Storage. EEL 5764: Graduate Computer Architecture. Storage Ann Gordon-Ross Electrical and Computer Engineering University of Florida.

CSE 502 Graduate Computer Architecture. Lec 22 Disk Storage

Computer Science 146. Computer Architecture

Review. EECS 252 Graduate Computer Architecture. Lec 18 Storage. Case for Storage. Outline

CS61C : Machine Structures

CS61C - Machine Structures. Week 7 - Disks. October 10, 2003 John Wawrzynek

Lecture 23: I/O Redundant Arrays of Inexpensive Disks Professor Randy H. Katz Computer Science 252 Spring 1996

Outline. EEL-4713 Computer Architecture I/O Systems. I/O System Design Issues. The Big Picture: Where are We Now? I/O Performance Measures

Today: Coda, xfs! Brief overview of other file systems. Distributed File System Requirements!

Topics. Lecture 8: Magnetic Disks

Agenda. Project #3, Part I. Agenda. Project #3, Part I. No SSE 12/1/10

Agenda. Agenda 11/12/12. Review - 6 Great Ideas in Computer Architecture

Appendix D: Storage Systems (Cont)

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Computer Architecture ECE 568

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Computer Architecture ECE 568

CS61C : Machine Structures

www-inst.eecs.berkeley.edu/~cs61c/

Chapter 6. Storage and Other I/O Topics

Page 1. Motivation: Who Cares About I/O? Lecture 5: I/O Introduction: Storage Devices & RAID. I/O Systems. Big Picture: Who cares about CPUs?

Lecture 23: Storage Systems. Topics: disk access, bus design, evaluation metrics, RAID (Sections )

RAID (Redundant Array of Inexpensive Disks) I/O Benchmarks ABCs of UNIX File Systems A Study Comparing UNIX File System Performance

Storage. Hwansoo Han

Dependency Redundancy & RAID

Concepts Introduced. I/O Cannot Be Ignored. Typical Collection of I/O Devices. I/O Issues

RAID (Redundant Array of Inexpensive Disks)

CS 61C: Great Ideas in Computer Architecture. Dependability. Guest Lecturer: Paul Ruan

UCB CS61C : Machine Structures

UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering. Computer Architecture ECE 568

Storage Systems. Storage Systems

I/O CANNOT BE IGNORED

Computer Organization and Structure. Bing-Yu Chen National Taiwan University

Computer Architecture Computer Science & Engineering. Chapter 6. Storage and Other I/O Topics BK TP.HCM

Lecture 25: Dependability and RAID

Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted from Hennessy & Patterson / 2003 Elsevier

CSE325 Principles of Operating Systems. Mass-Storage Systems. David P. Duggan. April 19, 2011

Appendix D: Storage Systems

High-Performance Storage Systems

5.11 Parallelism and Memory Hierarchy: Redundant Arrays of Inexpensive Disks 485.e1

Mass-Storage Structure

I/O CANNOT BE IGNORED

Administrivia. CMSC 411 Computer Systems Architecture Lecture 19 Storage Systems, cont. Disks (cont.) Disks - review

COS 318: Operating Systems. Storage Devices. Vivek Pai Computer Science Department Princeton University

Virtual Memory. Reading. Sections 5.4, 5.5, 5.6, 5.8, 5.10 (2) Lecture notes from MKP and S. Yalamanchili

Fault-Tolerant Storage and Implications for the Cloud Charles Snyder

Review. EECS 252 Graduate Computer Architecture. Lec 18 Storage. Introduction to Queueing Theory. Deriving Little s Law

Definition of RAID Levels

Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted from Hennessy & Patterson / 2003 Elsevier

CS152 Computer Architecture and Engineering Lecture 19: I/O Systems

CS61C : Machine Structures

Chapter 6 Storage and Other I/O Topics

Storage System COSC UCB

Chapter 10: Mass-Storage Systems

Storage systems. Computer Systems Architecture CMSC 411 Unit 6 Storage Systems. (Hard) Disks. Disk and Tape Technologies. Disks (cont.

Chapter 10: Mass-Storage Systems. Operating System Concepts 9 th Edition

I/O, Disks, and RAID Yi Shi Fall Xi an Jiaotong University

ECE Enterprise Storage Architecture. Fall 2018

- SLED: single large expensive disk - RAID: redundant array of (independent, inexpensive) disks

Page 1. Magnetic Disk Purpose Long term, nonvolatile storage Lowest level in the memory hierarchy. Typical Disk Access Time

Chapter 6. Storage and Other I/O Topics. ICE3003: Computer Architecture Fall 2012 Euiseong Seo

Where We Are in This Course Right Now. ECE 152 Introduction to Computer Architecture Input/Output (I/O) Copyright 2012 Daniel J. Sorin Duke University

Chapter 6. Storage and Other I/O Topics. Jiang Jiang

Recap of Networking Intro. Lecture #28 Networking & Disks Protocol Family Concept. Protocol for Networks of Networks?

Lecture 13. Storage, Network and Other Peripherals

Database Systems II. Secondary Storage

COS 318: Operating Systems. Storage Devices. Jaswinder Pal Singh Computer Science Department Princeton University

COS 318: Operating Systems. Storage Devices. Kai Li Computer Science Department Princeton University

Chapter 6. Storage and Other I/O Topics

HP AutoRAID (Lecture 5, cs262a)

HP AutoRAID (Lecture 5, cs262a)

Che-Wei Chang Department of Computer Science and Information Engineering, Chang Gung University

CS5460: Operating Systems Lecture 20: File System Reliability

Today s Papers. Array Reliability. RAID Basics (Two optional papers) EECS 262a Advanced Topics in Computer Systems Lecture 3

Mass-Storage. ICS332 - Fall 2017 Operating Systems. Henri Casanova

416 Distributed Systems. Errors and Failures Feb 1, 2016

An Introduction to RAID

Patterson s Projects, People, Impact

Filesystems Lecture 10. Credit: some slides by John Kubiatowicz and Anthony D. Joseph

u Covered: l Management of CPU & concurrency l Management of main memory & virtual memory u Currently --- Management of I/O devices

CMSC 424 Database design Lecture 12 Storage. Mihai Pop

CS 554: Advanced Database System

V. Mass Storage Systems

Modern RAID Technology. RAID Primer A Configuration Guide

Disks and RAID. CS 4410 Operating Systems. [R. Agarwal, L. Alvisi, A. Bracy, E. Sirer, R. Van Renesse]

Chapter 6. Storage and Other I/O Topics

Mass-Storage. ICS332 Operating Systems

CSE 153 Design of Operating Systems

Intelligent Drive Recovery (IDR): helping prevent media errors and disk failures with smart media scan

Mass-Storage Structure

CS370 Operating Systems

DATA DOMAIN INVULNERABILITY ARCHITECTURE: ENHANCING DATA INTEGRITY AND RECOVERABILITY

Distributed Video Systems Chapter 5 Issues in Video Storage and Retrieval Part 2 - Disk Array and RAID

Lecture for Storage System

PANASAS TIERED PARITY ARCHITECTURE

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Dependability and RAID. Great Ideas in Computer Architecture

Specifying Storage Servers for IP security applications

Computer Architecture 计算机体系结构. Lecture 6. Data Storage and I/O 第六讲 数据存储和输入输出. Chao Li, PhD. 李超博士

CS252 S05. CMSC 411 Computer Systems Architecture Lecture 18 Storage Systems 2. I/O performance measures. I/O performance measures

CISC 7310X. C11: Mass Storage. Hui Chen Department of Computer & Information Science CUNY Brooklyn College. 4/19/2018 CUNY Brooklyn College

Transcription:

Areal Density COSC 535 Advanced Computer Slides modified from Hennessy CS252 course slides Why do we need disks Magnetic Disks RAID Advanced Dependability/Reliability/Availability I/O Benchmarks, and Performance Conclusion COSC535 Advanced Computer /2/2 2 Shift in focus from computation to storage and communication of information Eg, Cray Research/Thinking Machines vs Google/Yahoo The Computing Revolution (96s to 98s) The Information Age (99 to today) Storage emphasizes reliability and scalability as well as cost-performance What is Software king that determines which HW features actually used? Operating System for storage Compiler for processor Also has own performance theory queuing theory balances throughput vs response time Figure D Cost versus access time for DRAM and magnetic disk in 98, 985, 99, 995, 2, and 25 The two-order-ofmagnitude gap in cost and five-order-of-magnitude gap in access times between semiconductor memory and rotating magnetic disks have inspired a host of competing technologies to try to fill them So far, such attempts have been made obsolete before production by improvements in magnetic disks, DRAMs, or both Note that between 99 and 25 the cost per gigabyte DRAM chips made less improvement, while disk cost made dramatic improvement COSC535 Advanced Computer /2/2 3 Copyright 2, Elsevier Inc All rights Reserved Why not store in DRAM They can be >,x faster But that performances costs 5X more per GB They have more bandwidth (2) About 8x, the bandwidth/gb is 2X But the $/bandwidth is 6x more costly Is there a technology cheaper than DRAM but faster than disks? Many have tried, but by the time they make it to market DRAM gets cheaper and disks get faster Right now SSDs are competing For servers, the M write limits is an issue, plus they are much more costly than disks Bits recorded along a track Metric is Bits Per Inch (BPI) Number of tracks per surface Metric is Tracks Per Inch (TPI) Disk Designs Brag about bit density per unit area Metric is Bits Per Square Inch: Areal Density = BPI x TPI Year Areal Density 973 2 979 8 989 63 997 3,9 2 7, 26 3, 2 636E+ Sep 2, Hitachi announced 636Gb/in 2,,,,, 97 98 99 2 2 Year COSC535 Advanced Computer /2/2 5 /2/2 6 NOW Handout Page S258 S99

956 IBM Ramac early 97s Winchester Developed for mainframe computers, proprietary interfaces Steady shrink in form factor: 27 in to 4 in Form factor and capacity drives market more than performance 97s developments 525 inch floppy disk formfactor (microcode into mainframe) Emergence of industry standard disk interfaces Early 98s: PCs and first generation workstations Mid 98s: Client/server computing Centralized storage on file server accelerates disk downsizing: 8 inch to 525 Mass market disk drives become a reality industry standards: SCSI, IPI, IDE 525 inch to 35 inch drives for PCs, End of proprietary interfaces 9s: Laptops => 25 inch drives 2s: What new devices leading to new drives? Continued advance in capacity (6%/yr) and bandwidth (4%/yr) Slow improvement in seek, rotation (8%/yr) Time to read whole disk Year Sequentially Randomly ( sector/seek) 99 4 minutes 6 hours 2 2 minutes week(!) 26 56 minutes 3 weeks (SCSI) 26 7 minutes 7 weeks (SATA) COSC535 Advanced Computer /2/2 7 COSC535 Advanced Computer /2/2 8 Katz and Patterson asked in 987: Can smaller disks be used to close gap in performance between disks and CPUs? Conventional: 4 disk designs 35 525 Low End 4 High End Disk Array: disk design 35 Low cost/mb High MB/volume High MB/watt Low cost/actuator Cost and Environmental Efficiencies COSC535 Advanced Computer COSC535 Advanced Computer /2/2 9 /2/2 Capacity Volume Power Data Rate I/O Rate MTTF Cost IBM 339K 2 GBytes 97 cu ft 3 KW 5 MB/s 6 I/Os/s 25 KHrs $25K IBM 35" 6 32 MBytes cu ft W 5 MB/s 55 I/Os/s 5 KHrs $2K x7 23 GBytes cu ft 9X KW 3X 2 MB/s 8X 39 IOs/s 6X??? Hrs $5K Disk Arrays have potential for large data and I/O rates, high MB per cu ft, high MB per KW, but what about reliability? COSC535 Advanced Computer COSC535 Advanced Computer /2/2 /2/2 2 Wiredcom Wikipediaorg NOW Handout Page 2 S258 S99 2

Reliability of N disks = Reliability of Disk N 5, Hours 7 disks = 7 hours Disk system MTTF: Drops from 6 years to month! Arrays (without redundancy) too unreliable to be useful! Hot spares support reconstruction in parallel with access: very high media availability can be achieved Files are "striped" across multiple disks Redundancy yields high data availability Availability: service still provided to user, even if some components failed Disks will still fail Contents reconstructed from data redundantly stored in the array Capacity penalty to store redundant info Bandwidth penalty to update redundant info COSC535 Advanced Computer /2/2 3 COSC535 Advanced Computer /2/2 4 JBOD Just use as a baseline Data may be striped or not striped recovery group Each disk is fully duplicated onto its mirror Very high availability can be achieved Bandwidth sacrifice on write: Logical write = two physical writes Reads may be optimized Most expensive solution: % capacity overhead (RAID 2 not interesting, so skip) COSC535 Advanced Computer COSC535 Advanced Computer /2/2 5 /2/2 6 How does mirroring interact with striping Say we had 8 disks and we wanted to use RAID on them Should we create 4 pairs of mirrored disks and stripe the data across them RAID + or RAID striped mirrors Should we create two sets of four disks each as RAID and mirror them RAID + or RAID mirrored stripes logical record Striped physical records COSC535 Advanced Computer COSC535 Advanced Computer /2/2 7 /2/2 8 P contains sum of other disks per stripe mod 2 ( parity ) If disk fails, subtract P from sum of other disks to find missing information P NOW Handout Page 3 S258 S99 3

Sum computed across recovery group to protect against hard disk failures, stored in P disk Logically, a single high capacity, high transfer rate disk: good for large transfers Wider arrays reduce capacity costs, but decreases availability 33% capacity cost for parity if 3 data disks and parity disk RAID 3 relies on parity disk to discover errors on Read But every sector has an error detection field To catch errors on read, rely on error detection field vs the parity disk Allows independent reads to different disks simultaneously COSC535 Advanced Computer /2/2 9 COSC535 Advanced Computer /2/2 2 Insides of 5 disks Example: small read D & D5, large write D2-D5 D D D2 D3 P D4 D5 D6 D7 P D8 D9 D D P D2 D3 D4 D5 D6 D7 D8 D9 D2 D2 D22 D23 P Disk Columns COSC535 Advanced Computer P P Increasing Logical Disk Address Stripe /2/2 2 RAID 4 works well for small reads Small writes (write to one disk): Option : read other data disks, create new sum and write to Parity Disk Option 2: since P has old sum, compare old data to new data, add the difference to P Small writes are limited by Parity Disk: Write to D, D5 both also write to P disk D D D2 D3 P D4 D5 D6 D7 P COSC535 Advanced Computer /2/2 22 Independent writes possible because of interleaved parity Example: write to D, D5 uses disks,, 3, 4 D D D2 D3 P D4 D5 D6 P D7 D8 D9 P D D D2 P D3 D4 D5 P D6 D7 D8 D9 D2 D2 D22 D23 P Disk Columns Increasing Logical Disk Addresses RAID-5: Small Write Algorithm D' new data Logical Write = 2 Physical Reads + 2 Physical Writes D D D2 D3 P + XOR COSC535 Advanced Computer COSC535 Advanced Computer /2/2 23 /2/2 24 old data ( Read) old (2 Read) parity + XOR (3 Write) (4 Write) D' D D2 D3 P' NOW Handout Page 4 S258 S99 4

Why > failure recovery? operator accidentally replaces the wrong disk during a failure since disk bandwidth is growing more slowly than disk capacity, the MTT Repair a disk in a RAID system is increasing increases the chances of a 2nd failure during repair since takes longer reading much more data during reconstruction meant increasing the chance of an uncorrectable media failure, which would result in data loss Network Appliance s row-diagonal parity or RAID-DP Like the standard RAID schemes, it uses redundant space based on parity calculation per stripe Since it is protecting against a double failure, it adds two check blocks per stripe of data If p+ disks total, p- disks have data; assume p=5 Row parity disk is just like in RAID 4 Even parity across the other 4 data blocks in its stripe Each block of the diagonal parity disk contains the even parity of the blocks in the same diagonal COSC535 Advanced Computer /2/2 25 COSC535 Advanced Computer /2/2 26 Row diagonal parity starts by recovering one of the 4 blocks on the failed disk using diagonal parity Since each diagonal misses one disk, and all diagonals miss a different disk, 2 diagonals are only missing block Once the data for those blocks is recovered, then the standard RAID recovery scheme can be used to recover two more blocks in the standard RAID 4 stripes Process continues until two failed disks are restored Row diagonal parity starts by recovering one of the 4 blocks on the failed disk using diagonal parity Since each diagonal misses one disk, and all diagonals miss a different disk, 2 diagonals are only missing block Once the data for those blocks is recovered, then the standard RAID recovery scheme can be used to recover two more blocks in the standard RAID 4 stripes Process continues until two failed disks are restored Data Disk Data Disk Data Disk 2 Data Disk 3 Row Parity Diagonal Parity 2 3 4 2 3 4 2 3 4 2 3 4 2 3 4 2 3 4 2 3 4 COSC535 Advanced Computer /2/2 27 COSC535 Advanced Computer /2/2 28 RAID-I (989) Consisted of a Sun 4/28 workstation with 28 MB of DRAM, four dual-string SCSI controllers, 28 525- inch SCSI disks and specialized disk striping software Today RAID is $24 billion dollar industry, 8% nonpc disks sold in RAIDs Disk Mirroring, Shadowing (RAID ) Each disk is fully duplicated onto its "shadow" Logical write = two physical writes % capacity overhead Parity Data Bandwidth Array (RAID 3) Parity computed horizontally Logically a single high data bw disk High I/O Rate Parity Array (RAID 5) Interleaved parity blocks Independent reads and writes Logical write = 2 reads + 2 writes COSC535 Advanced Computer COSC535 Advanced Computer /2/2 29 /2/2 3 NOW Handout Page 5 S258 S99 5

Examples on why precise definitions so important for reliability Is a programming mistake a fault, error, or failure? Are we talking about the time it was designed or the time the program is run? If the running program doesn t exercise the mistake, is it still a fault/error/failure? If an alpha particle hits a DRAM memory cell, is it a fault/error/failure if it doesn t change the value? Is it a fault/error/failure if the memory doesn t access the changed bit? Did a fault/error/failure still occur if the memory had error correction and delivered the corrected value to the CPU? Computer system dependability: quality of delivered service such that reliance can be placed on service Service is observed actual behavior as perceived by other system(s) interacting with this system s users Each module has ideal specified behavior, where service specification is agreed description of expected behavior A system failure occurs when the actual behavior deviates from the specified behavior Failure occurred due to an error, a defect in module The cause of an error is a fault When a fault occurs it creates a latent error, which becomes effective when it is activated When error actually affects the delivered service, a failure occurs (time from error to failure is error latency) COSC535 Advanced Computer /2/2 3 COSC535 Advanced Computer /2/2 32 An error is manifestation in the system of a fault, a failure is manifestation on the service of an error If an alpha particle hits a DRAM memory cell, is it a fault/error/failure if it doesn t change the value? Is it a fault/error/failure if the memory doesn t access the changed bit? Did a fault/error/failure still occur if the memory had error correction and delivered the corrected value to the CPU? An alpha particle hitting a DRAM can be a fault if it changes the memory, it creates an error error remains latent until effected memory word is read if the effected word error affects the delivered service, a failure occurs Hardware faults: Devices that fail, such alpha particle hitting a memory cell 2 Design faults: Faults in software (usually) and hardware design (occasionally) 3 Operation faults: Mistakes by operations and maintenance personnel 4 Environmental faults: Fire, flood, earthquake, power failure, and sabotage Also by duration: Transient faults exist for limited time and not recurring 2 Intermittent faults cause a system to oscillate between faulty and fault-free operation 3 Permanent faults do not correct themselves over time COSC535 Advanced Computer /2/2 33 COSC535 Advanced Computer /2/2 34 Fault-Tolerance (or more properly, Error-Tolerance): mask local faults (prevent errors from becoming failures) RAID disks Uninterruptible Power Supplies Cluster Failover Disaster Tolerance: masks site errors (prevent site errors from causing service failures) Protects against fire, flood, sabotage, Redundant system and service at remote site Use design diversity From Jim Gray s Talk at UC Berkeley on Fault Tolerance " /9/ COSC535 Advanced Computer /2/2 35 45 4 35 3 25 2 5 5 Mean Time to System Failure (years) by Cause environment 985 987 99 SOFTWARE 2 53 33 Years HARDWARE 29 9 3 Years MAINTENANCE 45 62 49 Years OPERATIONS 99 7 36 Years ENVIRONMENT 42 24 346 Years SYSTEM 8 2 2 Years Problem: Systematic Under-reporting total maintenance hardware operations software 985 987 989 From Jim Gray s Talk at UC Berkeley on Fault Tolerance " /9/ COSC535 Advanced Computer /2/2 36 NOW Handout Page 6 S258 S99 6

Rule of Thumb: Maintenance X HW so over 5 year product life, ~ 95% of cost is maintenance VAX crashes 85, 93 [Murp95]; extrap to Sys Man: N crashes/problem, SysAdmin action Actions: set params bad, bad config, bad app install HW/OS 7% in 85 to 28% in 93 In, %? COSC535 Advanced Computer /2/2 37 A cluster of 2 PCs in seven 7-foot high, 9-inch wide racks with 368 84 GB, 72 RPM, 35-inch IBM disks The PCs are P6-2MHz with 96 MB of DRAM each They run FreeBSD 3 and the hosts are connected via switched Mbit/second Ethernet Component Total in System Total Failed % Failed SCSI Controller 44 23% SCSI Cable 39 26% SCSI Disk 368 7 9% IDE Disk 24 6 25% Disk Enclosure -Backplane 46 3 283% Disk Enclosure - Power Supply 92 3 33% Ethernet Controller 2 5% Ethernet Switch 2 5% Ethernet Cable 42 23% CPU/Motherboard 2 % COSC535 Advanced Computer /2/2 38 Messages in system log for failed disk Hardware Failure (Peripheral device write fault [for] Field Replaceable Unit) Not Ready (Diagnostic failure: ASCQ = Component ID [of] Field Replaceable Unit) Recovered Error (Failure Prediction Threshold Exceeded [for] Field Replaceable Unit) Recovered Error (Failure Prediction Threshold Exceeded [for] Field Replaceable Unit) No log msgs Duration (hours) 763 86 46 9 33 5 43 7 System Type Unmanaged Managed Well Managed Fault Tolerant High-Availability Very-High-Availability Ultra-Availability Unavailable (min/year) 5, 5, 5 5 5 5 5 Availability 9% 99% 999% 9999% 99999% 999999% 9999999% Availability Class 2 3 4 5 6 7 UnAvailability = MTTR/MTBF can cut it in ½ by cutting MTTR or MTBF From Jim Gray s Talk at UC Berkeley on Fault Tolerance " /9/ /2/2 39 /2/2 4 HP claims HP-9 server HW and HP-UX OS can deliver 99999% availability guarantee in certain pre-defined, pre-tested customer environments Application faults? Operator faults? Environmental faults? Collocation sites (lots of computers in building on Internet) have network outage per year (~ day) power failure per year (~ day) Microsoft Network unavailable recently for a day due to problem in Domain Name Server: if only outage per year, 997% or 2 Nines Metrics: Response Time vs Throughput Proc 3 2 Queue Response Time (ms) % IOC Throughput (% total BW) Device % Response time = Queue + Device Service time COSC535 Advanced Computer COSC535 Advanced Computer /2/2 4 /2/2 42 NOW Handout Page 7 S258 S99 7

For better or worse, benchmarks shape a field Processor benchmarks classically aimed at response time for fixed sized problem I/O benchmarks typically measure throughput, possibly with upper limit on response times (or 9% of response times) Transaction Processing (TP) (or On-line TP=OLTP) If bank computer fails when customer withdraw money, TP system guarantees account debited if customer gets $ & account unchanged if no $ Airline reservation systems & banks use TP Atomic transactions makes this work Classic metric is Transactions Per Second (TPS) Goal: quantify variation in QoS metrics as events occur that affect system availability Leverage existing performance benchmarks to generate fair workloads to measure & trace quality of service metrics Use fault injection to compromise system hardware faults (disk, memory, network, power) software faults (corrupt input, driver error returns) maintenance events (repairs, SW/HW upgrades) Examine single-fault and multi-fault workloads the availability analogues of performance micro- and macro-benchmarks COSC535 Advanced Computer /2/2 43 COSC535 Advanced Computer /2/2 44 Linux Hits per second Solaris 22 25 2 25 2 95 9 6 4 2 8 Reconstruction 2 3 4 5 6 7 8 9 Reconstruction Hits/sec # failures tolerated 2 3 4 5 6 7 8 9 Time (minutes) Compares Linux and Solaris reconstruction Linux: minimal performance impact but longer window of vulnerability to second fault Solaris: large perf impact but restores redundancy fast COSC535 Advanced Computer 2 #failures tolerated 2 /2/2 45 Linux: favors performance over data availability automatically-initiated reconstruction, idle bandwidth virtually no performance impact on application very long window of vulnerability (>hr for 3GB RAID) Solaris: favors data availability over app perf automatically-initiated reconstruction at high BW as much as 34% drop in application performance short window of vulnerability ( minutes for 3GB) Windows: favors neither! manually-initiated reconstruction at moderate BW as much as 8% app performance drop somewhat short window of vulnerability (23 min/3gb) COSC535 Advanced Computer /2/2 46 Disks: Arial Density now 3%/yr vs %/yr in 2s Fault Latent errors in system Failure in service Components often fail slowly Real systems: problems in maintenance, operation as well as hardware, software COSC535 Advanced Computer /2/2 47 NOW Handout Page 8 S258 S99 8