First magnetic disks, the IBM 305 RAMAC (2 units shown) introduced in 1956. One platter shown top right. A RAMAC stored 5 million characters on 50 24-inch.

Slides:



Advertisements
Similar presentations
Redundant Array of Independent Disks (RAID) Striping of data across multiple media for expansion, performance and reliability.
Advertisements

A CASE FOR REDUNDANT ARRAYS OF INEXPENSIVE DISKS (RAID) D. A. Patterson, G. A. Gibson, R. H. Katz University of California, Berkeley.
Data Storage Solutions Module 1.2. Data Storage Solutions Upon completion of this module, you will be able to: List the common storage media and solutions.
CS224 Spring 2011 Computer Organization CS224 Chapter 6A: Disk Systems With thanks to M.J. Irwin, D. Patterson, and J. Hennessy for some lecture slide.
I/O Management and Disk Scheduling
Magnetic Disk Magnetic disks are the foundation of external memory on virtually all computer systems. A disk is a circular platter constructed of.
RAID Redundant Array of Independent Disks
CSCE430/830 Computer Architecture
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
CSE521: Introduction to Computer Architecture Mazin Yousif I/O Subsystem RAID (Redundant Array of Independent Disks)
Lecture 36: Chapter 6 Today’s topic –RAID 1. RAID Redundant Array of Inexpensive (Independent) Disks –Use multiple smaller disks (c.f. one large disk)
RAID Technology. Use Arrays of Small Disks? 14” 10”5.25”3.5” Disk Array: 1 disk design Conventional: 4 disk designs Low End High End Katz and Patterson.
CS 430 – Computer Architecture Disks
I/O Systems Processor Cache Memory - I/O Bus Main Memory I/O Controller Disk I/O Controller I/O Controller Graphics Network interrupts.
CS61C L40 I/O: Disks (1) Ho, Fall 2004 © UCB TA Casey Ho inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures Lecture 39 I/O : Disks Microsoft.
Disk Storage SystemsCSCE430/830 Disk Storage Systems CSCE430/830 Computer Architecture Lecturer: Prof. Hong Jiang Courtesy of Yifeng Zhu (U. Maine) Fall,
Computer ArchitectureFall 2007 © November 28, 2007 Karem A. Sakallah Lecture 24 Disk IO and RAID CS : Computer Architecture.
CS61C L16 Disks © UC Regents 1 CS61C - Machine Structures Lecture 16 - Disks October 20, 2000 David Patterson
1 Lecture 26: Storage Systems Topics: Storage Systems (Chapter 6), other innovations Final exam stats:  Highest: 95  Mean: 70, Median: 73  Toughest.
CS61C L40 I/O: Disks (1) Garcia, Fall 2004 © UCB Lecturer PSOE Dan Garcia inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures.
Computer ArchitectureFall 2008 © November 12, 2007 Nael Abu-Ghazaleh Lecture 24 Disk IO.
Disk Technologies. Magnetic Disks Purpose: – Long-term, nonvolatile, inexpensive storage for files – Large, inexpensive, slow level in the memory hierarchy.
Cs 61C L15 Disks.1 Patterson Spring 99 ©UCB CS61C Anatomy of I/O Devices: Magnetic Disks Lecture 15 March 10, 1999 Dave Patterson (http.cs.berkeley.edu/~patterson)
S.1 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output Cache Main Memory Secondary Memory (Disk)
DAP Fall.‘00 ©UCB 1 Storage Devices and RAID Professor David A. Patterson Computer Science 252 Fall 2000.
CS252/Patterson Lec 5.1 1/31/01 CS252 Graduate Computer Architecture Lecture 5: I/O Introduction: Storage Devices & RAID January 31, 2001 Prof. David A.
CS 61C L41 I/O Disks (1) Garcia, Spring 2004 © UCB Lecturer PSOE Dan Garcia inst.eecs.berkeley.edu/~cs61c CS61C : Machine.
CS252/Culler Lec 6.1 2/7/02 CS252 Graduate Computer Architecture I/O Introduction: Storage Devices & RAID Jason Hill.
RAID Systems CS Introduction to Operating Systems.
CS 61C: Great Ideas in Computer Architecture (Machine Structures) Lecture 39: IO Disks Instructor: Dan Garcia 1.
Lecture 11: Storage Systems Disk, RAID, Dependability Kai Bu
Redundant Array of Inexpensive Disks (RAID). Redundant Arrays of Disks Files are "striped" across multiple spindles Redundancy yields high data availability.
Lecture 11: Storage Systems Disk, RAID, Dependability Kai Bu
Chapter 6 RAID. Chapter 6 — Storage and Other I/O Topics — 2 RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f.
Eng. Mohammed Timraz Electronics & Communication Engineer University of Palestine Faculty of Engineering and Urban planning Software Engineering Department.
Storage & Peripherals Disks, Networks, and Other Devices.
Lecture 4 1 Reliability vs Availability Reliability: Is anything broken? Availability: Is the system still available to the user?
CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Storage Systems.
RAID Storage EEL4768, Fall 2009 Dr. Jun Wang Slides Prepared based on D&P Computer Architecture textbook, etc.
CPE 631: Storage Electrical and Computer Engineering University of Alabama in Huntsville Aleksandar Milenkovic
1 IKI20210 Pengantar Organisasi Komputer Kuliah No. 21: Peripheral 29 Nopember 2002 Bobby Nazief Johny Moningka
N-Tier Client/Server Architectures Chapter 4 Server - RAID Copyright 2002, Dr. Ken Hoganson All rights reserved. OS Kernel Concept RAID – Redundant Array.
I/O – Chapter 8 Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
1 Chapter 7: Storage Systems Introduction Magnetic disks Buses RAID: Redundant Arrays of Inexpensive Disks.
Lecture 9 of Advanced Databases Storage and File Structure (Part II) Instructor: Mr.Ahmed Al Astal.
Redundant Array of Inexpensive Disks aka Redundant Array of Independent Disks (RAID) Modified from CCT slides.
Eng. Mohammed Timraz Electronics & Communication Engineer University of Palestine Faculty of Engineering and Urban planning Software Engineering Department.
Storage. 10/20/20152 Case for Storage Shift in focus from computation to communication and storage of information –E.g., Cray Research/Thinking Machines.
CS 136, Advanced Architecture Storage. CS136 2 Case for Storage Shift in focus from computation to communication and storage of information –E.g., Cray.
Disk Storage SystemsCSCE430/830 Disk Storage Systems CSCE430/830 Computer Architecture Lecturer: Prof. Hong Jiang Courtesy of Yifeng Zhu (U. Maine) Fall,
Αρχιτεκτονική Υπολογιστών Ενότητα # 6: RAID Διδάσκων: Γεώργιος Κ. Πολύζος Τμήμα: Πληροφορικής.
CS 61C: Great Ideas in Computer Architecture Dependability and RAID Lecture 25 Instructors: John Wawrzynek & Vladimir Stojanovic
Chapter 8 Secondary Memory. Topics Types of External Memory Magnetic Disk Optical Magnetic Tape.
1 Lecture 27: Disks Today’s topics:  Disk basics  RAID  Research topics.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
W4118 Operating Systems Instructor: Junfeng Yang.
Lecture 11: Storage Systems Disk, RAID, Dependability Kai Bu
CMSC 611: Advanced Computer Architecture I/O & Storage Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
I/O Errors 1 Computer Organization II © McQuain RAID Redundant Array of Inexpensive (Independent) Disks – Use multiple smaller disks (c.f.
CS Introduction to Operating Systems
Multiple Platters.
RAID Redundant Arrays of Independent Disks
Module: Storage Systems
Vladimir Stojanovic & Nicholas Weaver
IT 251 Computer Organization and Architecture
CS 554: Advanced Database System Notes 02: Hardware
Appendix D– Storage Systems
CS252 Graduate Computer Architecture I/O Introduction: Storage Devices & RAID Jason Hill.
Presentation transcript:

First magnetic disks, the IBM 305 RAMAC (2 units shown) introduced in One platter shown top right. A RAMAC stored 5 million characters on inch diameter platters. Two access arms moved up and down to select a disk, and in/out to select a track. Right: a variety of disk drives: 8”, 5.25”, 3.5”, 1.8” and 1”.

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Storage Anselmo Lastra

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 3 Outline Magnetic Disks RAID Advanced Dependability/Reliability/Availability I/O Benchmarks, Performance and Dependability Conclusion

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 4 Disk Figure of Merit: Areal Density Bits recorded along a track ♦ Metric is Bits Per Inch (BPI) Number of tracks per surface ♦ Metric is Tracks Per Inch (TPI) Disk metric is bit density per unit area ♦ Metric is Bits Per Square Inch: Areal Density = BPI x TPI

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 5 Historical Perspective 1956 IBM Ramac — early 1970s Winchester ♦ For mainframe computers, proprietary interfaces ♦ Steady shrink in form factor: 27 in. to 14 in. Form factor and capacity drives market more than performance 1970s developments ♦ 5.25 inch floppy disk form factor ♦ Emergence of industry standard disk interfaces Early 1980s: PCs and first generation workstations Mid 1980s: Client/server computing ♦ Centralized storage on file server ♦ Disk downsizing: 8 inch to 5.25 ♦ Mass market disk drives become a reality industry standards: SCSI, IPI, IDE 5.25 inch to 3.5 inch drives for PCs; End of proprietary interfaces 1990s: Laptops => 2.5 inch drives 2000s: 1.8” used in media players (1” microdrive didn’t do as well)

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 6 Current Disks Caches to hold recently accessed blocks Microprocessor and command buffer to enable reordering of accesses

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 7 Future Disk Size and Performance Continued advance in capacity (60%/yr) and bandwidth (40%/yr) Slow improvement in seek, rotation (8%/yr) Time to read whole disk YearSequentiallyRandomly (1 sector/seek) minutes6 hours minutes 1 week(!) minutes 3 weeks (SCSI) minutes 7 weeks (SATA) Cost has dropped by 100,000 since 1983

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 8 Arrays of Small Disks? 14” 10”5.25”3.5” Disk Array: 1 disk design Conventional: 4 disk designs Low End High End Katz and Patterson asked in 1987: Can smaller disks be used to close gap in performance between disks and CPUs?

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 9 Advantages of Small Form Factor Disk Drives Low cost/MB High MB/volume High MB/watt Low cost/Actuator Cost and Environmental Efficiencies

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 10 Replace Small Number of Large Disks with Large Number of Small Disks! (1988 Disks) Capacity Volume Power Data Rate I/O Rate MTTF Cost IBM 3390K 20 GBytes 97 cu. ft. 3 KW 15 MB/s 600 I/Os/s 250 KHrs $250K IBM 3.5" MBytes 0.1 cu. ft. 11 W 1.5 MB/s 55 I/Os/s 50 KHrs $2K x70 23 GBytes 11 cu. ft. 1 KW 120 MB/s 3900 IOs/s ??? Hrs $150K Disk Arrays have potential for large data and I/O rates, high MB per cu. ft., high MB per KW, but what about reliability? 9X 3X 8X 6X

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 11 Array Reliability Reliability of N disks = Reliability of 1 Disk ÷ N 50,000 Hours ÷ 70 disks = 700 hours Disk system MTTF: Drops from 6 years to 1 month! Arrays (without redundancy) too unreliable to be useful! Hot spares support reconstruction in parallel with access: very high media availability can be achieved Hot spares support reconstruction in parallel with access: very high media availability can be achieved

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 12 Redundant Arrays of (Inexpensive) Disks Files are "striped" across multiple disks Redundancy yields high data availability ♦ Availability: service still provided to user, even if some components failed Disks will still fail Contents reconstructed from data redundantly stored in the array  Capacity penalty to store redundant info  Bandwidth penalty to update redundant info

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 13 RAID0 Performance only No redundancy Stripe data to get higher bandwidth Latency not improved

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 14 Redundant Arrays of Inexpensive Disks RAID 1: Disk Mirroring/Shadowing Each disk is fully duplicated onto its “mirror” Very high availability can be achieved Bandwidth sacrifice on write: Logical write = two physical writes Reads may be optimized Most expensive solution: 100% capacity overhead ( RAID 2 not interesting, so skip) recovery group

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 15 Redundant Array of Inexpensive Disks RAID 3: Parity Disk P logical record P contains sum of other disks per stripe mod 2 (“parity”) If disk fails, subtract P from sum of other disks to find missing information Striped physical records

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 16 RAID 3 Sum computed across recovery group to protect against hard disk failures, stored in P disk Logically, a single high capacity, high transfer rate disk: good for large transfers Wider arrays reduce capacity costs, but decreases availability 33% capacity cost for parity if 3 data disks and 1 parity disk

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 17 Inspiration for RAID 4 RAID 3 relies on parity disk to discover errors on Read But every sector has an error detection field To catch errors on read, rely on error detection field vs. the parity disk Allows independent reads to different disks simultaneously

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 18 Redundant Arrays of Inexpensive Disks RAID 4: High I/O Rate Parity D0D1D2 D3 P D4D5D6 PD7 D8D9 PD10 D11 D12 PD13 D14 D15 P D16D17 D18 D19 D20D21D22 D23 P Disk Columns Increasing Logical Disk Address Stripe Insides of 5 disks Example: small read D0 & D5, large write D12-D15 Example: small read D0 & D5, large write D12-D15

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 19 Inspiration for RAID 5 RAID 4 works well for small reads Small writes (write to one disk): ♦ Option 1: read other data disks, create new sum and write to Parity Disk ♦ Option 2: since P has old sum, compare old data to new data, add the difference to P (2 reads, 2 writes) Small writes are limited by Parity Disk: Write to D0, D5 both also write to P disk D0 D1D2 D3 P D4 D5 D6 P D7

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 20 Redundant Arrays of Inexpensive Disks RAID 5: High I/O Rate Interleaved Parity Independent writes possible because of interleaved parity Independent writes possible because of interleaved parity D0D1D2 D3 P D4D5D6 P D7 D8D9P D10 D11 D12PD13 D14 D15 PD16D17 D18 D19 D20D21D22 D23 P Disk Columns Increasing Logical Disk Addresses Example: write to D0, D5 uses disks 0, 1, 3, 4

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 21 Downside of Disk Arrays: Cost of Small Writes D0D1D2 D3 P D0' + + D1D2 D3 P' new data old data old parity XOR (1. Read) (2. Read) (3. Write) (4. Write) RAID-5: Small Write Algorithm 1 Logical Write = 2 Physical Reads + 2 Physical Writes

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 22 RAID 6: Recovering from 2 failures Like the standard RAID schemes, it uses redundant space based on parity calculation per stripe Idea is that operator may make mistake and swap wrong disk, or 2 nd disk may fail while replacing 1 st Since it is protecting against a double failure, it adds two check blocks per stripe of data. ♦ If p+1 disks total, p-1 disks have data; assume p=5

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 23 Summary of RAID Techniques Disk Mirroring, Shadowing (RAID 1) Each disk is fully duplicated onto its "shadow" Logical write = two physical writes 100% capacity overhead Parity Data Bandwidth Array (RAID 3) Parity computed horizontally Logically a single high data bw disk High I/O Rate Parity Array (RAID 5) Interleaved parity blocks Independent reads and writes Logical write = 2 reads + 2 writes

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 24 HW Failures in Real Systems: Tertiary Disks A cluster of 20 PCs in seven 7-foot high, 19-inch wide racks with GB, 7200 RPM, 3.5-inch IBM disks. The PCs are P6-200MHz with 96 MB of DRAM each.

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 25 Does Hardware Fail Fast? 4 disks that failed in Tertiary Disk Author says that almost all disk failures began as transient failures. The operator had to decide when to replace.

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 26 Internet Archive Good section in text about Internet Archive ♦ In 2006, over a petabyte of disk (10 15 bytes) ♦ Growing at 20 terabytes (10 12 ) per month ♦ Now says ~ 3PB Each PC was 1GHz Via, 512MB, dissipates 80 W Each node had GB drives 40 nodes/rack ♦ Petabyte takes 12 racks PC cost $500, each disk $375, 40-port Ethernet switch $3000

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Capricorn PS now AMD Athlon 64 (x2) 4 SATA disks (1-4TB) 92 Watts/node 40 nodes/rack So 160 TB/rack 24 kW/petabyte ♦ So 576 kWh/day, ~17,000/mo ♦ Ave. U.S. house used 920 kWh a month in 2006 ♦ Best housed in KY? 27

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Drives Today Just ordered simple RAID enclosure (just levels 0, 1) ♦ $63 Two 1TB SATA drives ♦ $85/ea 28

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL 29 Summary Disks: Areal Density now 30%/yr vs. 100%/yr in 2000s Components often fail slowly Real systems: problems in maintenance, operation as well as hardware, software