1 Chapter 7: Storage Systems Introduction Magnetic disks Buses RAID: Redundant Arrays of Inexpensive Disks.

Slides:



Advertisements
Similar presentations
RAID Redundant Arrays of Independent Disks Courtesy of Satya, Fall 99.
Advertisements

Redundant Array of Independent Disks (RAID) Striping of data across multiple media for expansion, performance and reliability.
1 Lecture 18: RAID n I/O bottleneck n JBOD and SLED n striping and mirroring n classic RAID levels: 1 – 5 n additional RAID levels: 6, 0+1, 10 n RAID usage.
Faculty of Information Technology Department of Computer Science Computer Organization Chapter 7 External Memory Mohammad Sharaf.
RAID (Redundant Arrays of Independent Disks). Disk organization technique that manages a large number of disks, providing a view of a single disk of High.
RAID A RRAYS Redundant Array of Inexpensive Discs.
Triple-Parity RAID and Beyond Hai Lu. RAID RAID, an acronym for redundant array of independent disks or also known as redundant array of inexpensive disks,
RAID Redundant Array of Independent Disks
CSCE430/830 Computer Architecture
CA 714CA Midterm Review. C5 Cache Optimization Reduce miss penalty –Hardware and software Reduce miss rate –Hardware and software Reduce hit time –Hardware.
Operating Systems ECE344 Ashvin Goel ECE University of Toronto Disks and RAID.
CSE521: Introduction to Computer Architecture Mazin Yousif I/O Subsystem RAID (Redundant Array of Independent Disks)
Lecture 36: Chapter 6 Today’s topic –RAID 1. RAID Redundant Array of Inexpensive (Independent) Disks –Use multiple smaller disks (c.f. one large disk)
CSCE 212 Chapter 8 Storage, Networks, and Other Peripherals Instructor: Jason D. Bakos.
RAID Technology. Use Arrays of Small Disks? 14” 10”5.25”3.5” Disk Array: 1 disk design Conventional: 4 disk designs Low End High End Katz and Patterson.
REDUNDANT ARRAY OF INEXPENSIVE DISCS RAID. What is RAID ? RAID is an acronym for Redundant Array of Independent Drives (or Disks), also known as Redundant.
EECC551 - Shaaban #1 Lec # 13 Winter Magnetic Disk CharacteristicsMagnetic Disk Characteristics I/O Connection StructureI/O Connection Structure.
Computer ArchitectureFall 2007 © November 28, 2007 Karem A. Sakallah Lecture 24 Disk IO and RAID CS : Computer Architecture.
Other Disk Details. 2 Disk Formatting After manufacturing disk has no information –Is stack of platters coated with magnetizable metal oxide Before use,
1 Lecture 26: Storage Systems Topics: Storage Systems (Chapter 6), other innovations Final exam stats:  Highest: 95  Mean: 70, Median: 73  Toughest.
First magnetic disks, the IBM 305 RAMAC (2 units shown) introduced in One platter shown top right. A RAMAC stored 5 million characters on inch.
Lecture 3: A Case for RAID (Part 1) Prof. Shahram Ghandeharizadeh Computer Science Department University of Southern California.
Computer ArchitectureFall 2008 © November 12, 2007 Nael Abu-Ghazaleh Lecture 24 Disk IO.
1 Lecture 23: Multiprocessors Today’s topics:  RAID  Multiprocessor taxonomy  Snooping-based cache coherence protocol.
S.1 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output Cache Main Memory Secondary Memory (Disk)
RAID Systems CS Introduction to Operating Systems.
Servers Redundant Array of Inexpensive Disks (RAID) –A group of hard disks is called a disk array FIGURE Server with redundant NICs.
Lecture 11: Storage Systems Disk, RAID, Dependability Kai Bu
Redundant Array of Inexpensive Disks (RAID). Redundant Arrays of Disks Files are "striped" across multiple spindles Redundancy yields high data availability.
ICOM 6005 – Database Management Systems Design Dr. Manuel Rodríguez-Martínez Electrical and Computer Engineering Department Lecture 6 – RAID ©Manuel Rodriguez.
Lecture 11: Storage Systems Disk, RAID, Dependability Kai Bu
Chapter 6 RAID. Chapter 6 — Storage and Other I/O Topics — 2 RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f.
Storage & Peripherals Disks, Networks, and Other Devices.
Lecture 4 1 Reliability vs Availability Reliability: Is anything broken? Availability: Is the system still available to the user?
CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Storage Systems.
CSE 321b Computer Organization (2) تنظيم الحاسب (2) 3 rd year, Computer Engineering Winter 2015 Lecture #4 Dr. Hazem Ibrahim Shehata Dept. of Computer.
RAID Storage EEL4768, Fall 2009 Dr. Jun Wang Slides Prepared based on D&P Computer Architecture textbook, etc.
L/O/G/O External Memory Chapter 3 (C) CS.216 Computer Architecture and Organization.
N-Tier Client/Server Architectures Chapter 4 Server - RAID Copyright 2002, Dr. Ken Hoganson All rights reserved. OS Kernel Concept RAID – Redundant Array.
I/O – Chapter 8 Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
Disk Access. DISK STRUCTURE Sector: Smallest unit of data transfer from/to disk; 512B 2/4/8 adjacent sectors transferred together: Blocks Read/write heads.
RAID COP 5611 Advanced Operating Systems Adapted from Andy Wang’s slides at FSU.
Lecture 9 of Advanced Databases Storage and File Structure (Part II) Instructor: Mr.Ahmed Al Astal.
Redundant Array of Inexpensive Disks aka Redundant Array of Independent Disks (RAID) Modified from CCT slides.
RAID REDUNDANT ARRAY OF INEXPENSIVE DISKS. Why RAID?
Disks Chapter 5 Thursday, April 5, Today’s Schedule Input/Output – Disks (Chapter 5.4)  Magnetic vs. Optical Disks  RAID levels and functions.
"1"1 Introduction to Managing Data " Describe problems associated with managing large numbers of disks " List requirements for easily managing large amounts.
Chapter 12 – Mass Storage Structures (Pgs )
Αρχιτεκτονική Υπολογιστών Ενότητα # 6: RAID Διδάσκων: Γεώργιος Κ. Πολύζος Τμήμα: Πληροφορικής.
Storage and File structure COP 4720 Lecture 20 Lecture Notes.
Chapter 8 Secondary Memory. Topics Types of External Memory Magnetic Disk Optical Magnetic Tape.
CS 6290 I/O and Storage Milos Prvulovic. Storage Systems I/O performance (bandwidth, latency) –Bandwidth improving, but not as fast as CPU –Latency improving.
1 Lecture 27: Disks Today’s topics:  Disk basics  RAID  Research topics.
1 Lecture 23: Storage Systems Topics: disk access, bus design, evaluation metrics, RAID (Sections )
Chapter 14: Mass-Storage Systems Disk Structure. Disk Scheduling. RAID.
W4118 Operating Systems Instructor: Junfeng Yang.
LECTURE 13 I/O. I/O CANNOT BE IGNORED Assume a program requires 100 seconds, 90 seconds for main memory, 10 seconds for I/O. Assume main memory access.
Lecture 11: Storage Systems Disk, RAID, Dependability Kai Bu
CMSC 611: Advanced Computer Architecture I/O & Storage Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
Silberschatz, Galvin and Gagne ©2013 Operating System Concepts – 9 th Edition Chapter 10: Mass-Storage Systems.
I/O Errors 1 Computer Organization II © McQuain RAID Redundant Array of Inexpensive (Independent) Disks – Use multiple smaller disks (c.f.
CS Introduction to Operating Systems
Multiple Platters.
RAID Redundant Arrays of Independent Disks
Vladimir Stojanovic & Nicholas Weaver
Lecture 13 I/O.
RAID RAID Mukesh N Tekwani
Overview Continuation from Monday (File system implementation)
TECHNICAL SEMINAR PRESENTATION
RAID RAID Mukesh N Tekwani April 23, 2019
Presentation transcript:

1 Chapter 7: Storage Systems Introduction Magnetic disks Buses RAID: Redundant Arrays of Inexpensive Disks

2 I/O Performance Amdahl's Law: Continuing improve only CPU performance Performance is not the only concern –Reliability –Availability –Dependenability –Serviceability CPUI/OOverall Improvement ~ ~10

3 Magnetic Disks Average Access Time: (mostly due to seek and rotation) = Average Seek Time + Ave. Rotation Delay + Transfer Time + Controller Delay Areal Density = Tracks/Inch on disk surface * Bits/Inch on track –Increase beyond Moore’s Law lately –<$1 per gigabyte today Cost vs. Access Time: Still huge gap among SRAM, DRAM, and magnetic disks –Technology to fill the gap? Other Technology –Optical disks, Flash memory

4 Technology Trend Component –IC technology: transistor increases 55% per year –DRAM: density increases 40-60% per year –Disk: density increases 100% per year lately –Network: Ethernet from 10->100Mb for 10 years; 100Mb->1Gb for 5 years DRAM/Disk:

5 Buses Shared communication links between subsystems: CPU bus, I/O bus, MP System bus etc. Bus design considerations –Bus physics: Driver design, flight-time, reflection, skew, glitches, cross talk, etc. –Bus width, separated or combined address / data buses. –Multiple bus masters and bus arbitration mechanism (must be fair and dead-lock free) –Simple bus (non-pipelined) vs split-transaction bus (pipelined) –Synchronous vs asynchronous buses –Multiprocessor bus: May include cache coherence control protocol (snooping bus)

6 RAID RAID 0: Striping across a set of disks makes collection appears as a single large disk, but no redundancy RAID 1: Mirroring; Maintain two copies, when one fails, goes to the backup Combined RAID 0, 1 –RAID10: striped mirrows –RAID01: mirrored stripes RAID 2: Memory-style ECC (not used) RAID 3: Bit-Interleaved Parity; Keep Parity bit in redundant disk to recover when single failure Mirror is a special case with one parity per bit

7 RAID (cont.) RAID 4 and 5:RAID 4 and 5: Block Interleaved Parity and Distributed Block Interleaved Parity RAID4 RAID5 Disk 0 Disk 1 Disk 2 Disk 3 Disk P0 4567P P P P4 Disk 0 Disk 1 Disk 2 Disk 3 Disk P0 456P17 89P P P

8 Assume 4 data disks, D 0, D 1, D 2, D 3 and one parity disk P For RAID 3, an small update of D 0 requires: For RAID 4&5, a small update of D 0 only requires Small Update: RAID 3 vs. RAID 4&5

9 Inspiration for RAID 5 RAID 4 works well for small reads Small writes (write to one disk): –Option 1: read other data disks, create new sum and write to Parity Disk –Option 2: since P has old sum, compare old data to new data, add the difference to P Small writes are limited by Parity Disk: Write to D0, D5 both also write to P disk D0 D1D2 D3 P D4 D5 D6 P D7

10 Redundant Arrays of Inexpensive Disks RAID 5: High I/O Rate Interleaved Parity Independent writes possible because of interleaved parity Independent writes possible because of interleaved parity D0D1D2 D3 P D4D5D6 P D7 D8D9P D10 D11 D12PD13 D14 D15 PD16D17 D18 D19 D20D21D22 D23 P Disk Columns Increasing Logical Disk Addresses Example: write to D0, D5 uses disks 0, 1, 3, 4

11 Problems of Disk Arrays: Small Writes D0D1D2 D3 P D0' + + D1D2 D3 P' new data old data old parity XOR (1. Read) (2. Read) (3. Write) (4. Write) RAID-5: Small Write Algorithm 1 Logical Write = 2 Physical Reads + 2 Physical Writes

12 RAID 6: Recovering from 2 failures Why > 1 failure recovery? –operator accidentally replaces the wrong disk during a failure –since disk bandwidth is growing more slowly than disk capacity, the MTT Repair a disk in a RAID system is increasing  increases the chances of a 2nd failure during repair since takes longer –reading much more data during reconstruction meant increasing the chance of an uncorrectable media failure, which would result in data loss

13 RAID 6: Recovering from 2 failures Network Appliance’s row-diagonal parity or RAID- DP Like the standard RAID schemes, it uses redundant space based on parity calculation per stripe Since it is protecting against a double failure, it adds two check blocks per stripe of data. –If p+1 disks total, p-1 disks have data; assume p=5 Row parity disk is just like in RAID 4 –Even parity across the other 4 data blocks in its stripe Each block of the diagonal parity disk contains the even parity of the blocks in the same diagonal

14 Example p = 5 Row diagonal parity starts by recovering one of the 4 blocks on the failed disk using diagonal parity –Since each diagonal misses one disk, and all diagonals miss a different disk, 2 diagonals are only missing 1 block Once the data for those blocks is recovered, then the standard RAID recovery scheme can be used to recover two more blocks in the standard RAID 4 stripes Process continues until two failed disks are restored Data Disk 0 Data Disk 1 Data Disk 2 Data Disk 3 Row Parity Diagona l Parity

15 Summary: RAID Techniques: Goal was performance, popularity due to reliability of storage Disk Mirroring, Shadowing (RAID 1) Each disk is fully duplicated onto its "shadow" Logical write = two physical writes 100% capacity overhead Parity Data Bandwidth Array (RAID 3) Parity computed horizontally Logically a single high data bw disk High I/O Rate Parity Array (RAID 5) Interleaved parity blocks Independent reads and writes Logical write = 2 reads + 2 writes