CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Storage Systems.

Slides:



Advertisements
Similar presentations
A Case for Redundant Arrays Of Inexpensive Disks Paper By David A Patterson Garth Gibson Randy H Katz University of California Berkeley.
Advertisements

Disk Arrays COEN 180. Large Storage Systems Collection of disks to store large amount of data. Performance advantage: Each drive can satisfy only so many.
Redundant Array of Independent Disks (RAID) Striping of data across multiple media for expansion, performance and reliability.
I/O Management and Disk Scheduling
1 Lecture 18: RAID n I/O bottleneck n JBOD and SLED n striping and mirroring n classic RAID levels: 1 – 5 n additional RAID levels: 6, 0+1, 10 n RAID usage.
CS 346 – April 4 Mass storage –Disk formatting –Managing swap space –RAID Commitment –Please finish chapter 12.
Faculty of Information Technology Department of Computer Science Computer Organization Chapter 7 External Memory Mohammad Sharaf.
RAID (Redundant Arrays of Independent Disks). Disk organization technique that manages a large number of disks, providing a view of a single disk of High.
RAID Oh yes Whats RAID? Redundant Array (of) Independent Disks. A scheme involving multiple disks which replicates data across multiple drives. Methods.
Magnetic Disk Magnetic disks are the foundation of external memory on virtually all computer systems. A disk is a circular platter constructed of.
RAID Redundant Array of Independent Disks
CS 6560: Operating Systems Design
CSCE430/830 Computer Architecture
Operating Systems ECE344 Ashvin Goel ECE University of Toronto Disks and RAID.
Chapter 6 External Memory Disk and RAID (Redundant Arrays of Independent Disks) CS-147 Fall 2010 Jonathan Wang.
RAID- Redundant Array of Inexpensive Drives. Purpose Provide faster data access and larger storage Provide data redundancy.
R.A.I.D. Copyright © 2005 by James Hug Redundant Array of Independent (or Inexpensive) Disks.
Lecture 36: Chapter 6 Today’s topic –RAID 1. RAID Redundant Array of Inexpensive (Independent) Disks –Use multiple smaller disks (c.f. one large disk)
CSCE 212 Chapter 8 Storage, Networks, and Other Peripherals Instructor: Jason D. Bakos.
REDUNDANT ARRAY OF INEXPENSIVE DISCS RAID. What is RAID ? RAID is an acronym for Redundant Array of Independent Drives (or Disks), also known as Redundant.
Computer ArchitectureFall 2007 © November 28, 2007 Karem A. Sakallah Lecture 24 Disk IO and RAID CS : Computer Architecture.
Other Disk Details. 2 Disk Formatting After manufacturing disk has no information –Is stack of platters coated with magnetizable metal oxide Before use,
1 Lecture 26: Storage Systems Topics: Storage Systems (Chapter 6), other innovations Final exam stats:  Highest: 95  Mean: 70, Median: 73  Toughest.
Lecture 3: A Case for RAID (Part 1) Prof. Shahram Ghandeharizadeh Computer Science Department University of Southern California.
Computer ArchitectureFall 2008 © November 12, 2007 Nael Abu-Ghazaleh Lecture 24 Disk IO.
Disks CS 416: Operating Systems Design, Spring 2001 Department of Computer Science Rutgers University
RAID Systems CS Introduction to Operating Systems.
CSE 451: Operating Systems Winter 2010 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura.
Transactions and Reliability. File system components Disk management Naming Reliability  What are the reliability issues in file systems? Security.
ICOM 6005 – Database Management Systems Design Dr. Manuel Rodríguez-Martínez Electrical and Computer Engineering Department Lecture 6 – RAID ©Manuel Rodriguez.
Chapter 6 RAID. Chapter 6 — Storage and Other I/O Topics — 2 RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f.
Storage & Peripherals Disks, Networks, and Other Devices.
1 Recitation 8 Disk & File System. 2 Disk Scheduling Disks are at least four orders of magnitude slower than main memory –The performance of disk I/O.
CSE 321b Computer Organization (2) تنظيم الحاسب (2) 3 rd year, Computer Engineering Winter 2015 Lecture #4 Dr. Hazem Ibrahim Shehata Dept. of Computer.
L/O/G/O External Memory Chapter 3 (C) CS.216 Computer Architecture and Organization.
N-Tier Client/Server Architectures Chapter 4 Server - RAID Copyright 2002, Dr. Ken Hoganson All rights reserved. OS Kernel Concept RAID – Redundant Array.
I/O – Chapter 8 Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
1 Chapter 7: Storage Systems Introduction Magnetic disks Buses RAID: Redundant Arrays of Inexpensive Disks.
RAID COP 5611 Advanced Operating Systems Adapted from Andy Wang’s slides at FSU.
Lecture 9 of Advanced Databases Storage and File Structure (Part II) Instructor: Mr.Ahmed Al Astal.
Parity Logging O vercoming the Small Write Problem in Redundant Disk Arrays Daniel Stodolsky Garth Gibson Mark Holland.
Redundant Array of Inexpensive Disks aka Redundant Array of Independent Disks (RAID) Modified from CCT slides.
RAID SECTION (2.3.5) ASHLEY BAILEY SEYEDFARAZ YASROBI GOKUL SHANKAR.
Redundant Array of Independent Disks.  Many systems today need to store many terabytes of data.  Don’t want to use single, large disk  too expensive.
RAID Disk Arrays Hank Levy. 212/5/2015 Basic Problems Disks are improving, but much less fast than CPUs We can use multiple disks for improving performance.
RAID Systems Ver.2.0 Jan 09, 2005 Syam. RAID Primer Redundant Array of Inexpensive Disks random, real-time, redundant, array, assembly, interconnected,
Auxiliary Memory Magnetic Disk:
1 Lecture 27: Disks Today’s topics:  Disk basics  RAID  Research topics.
W4118 Operating Systems Instructor: Junfeng Yang.
RAID TECHNOLOGY RASHMI ACHARYA CSE(A) RG NO
Network-Attached Storage. Network-attached storage devices Attached to a local area network, generally an Ethernet-based network environment.
Silberschatz, Galvin and Gagne ©2013 Operating System Concepts – 9 th Edition Chapter 10: Mass-Storage Systems.
I/O Errors 1 Computer Organization II © McQuain RAID Redundant Array of Inexpensive (Independent) Disks – Use multiple smaller disks (c.f.
CSE 451: Operating Systems Spring 2010 Module 18 Redundant Arrays of Inexpensive Disks (RAID) John Zahorjan Allen Center 534.
CS Introduction to Operating Systems
CMPE Database Systems Workshop June 16 Class Meeting
A Case for Redundant Arrays of Inexpensive Disks (RAID) -1988
A Case for Redundant Arrays of Inexpensive Disks (RAID)
Transactions and Reliability
Multiple Platters.
Disks and RAID.
Vladimir Stojanovic & Nicholas Weaver
CSE 451: Operating Systems Spring 2006 Module 18 Redundant Arrays of Inexpensive Disks (RAID) John Zahorjan Allen Center.
RAID RAID Mukesh N Tekwani
ICOM 6005 – Database Management Systems Design
TECHNICAL SEMINAR PRESENTATION
RAID Redundant Array of Inexpensive (Independent) Disks
UNIT IV RAID.
RAID RAID Mukesh N Tekwani April 23, 2019
IT 344: Operating Systems Winter 2007 Module 18 Redundant Arrays of Inexpensive Disks (RAID) Chia-Chi Teng CTB
Presentation transcript:

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Storage Systems

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Why Worry About Storage Systems Response time = CPU time + I/O time Suppose 95% of work is done in the CPU, 5% is I/O –If the CPU is improved by a factor of 100, what is the Speedup? –Sp = 1 / ( (0.95 / 100)) = 16.8 –83.2% of the improvement is squandered due to poor I/O Future performance gains must consider faster I/O Amdahl’s Law  All parts of a system need to be improved somewhat equally

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Magnetic Disk Physically consist of –Tracks –Sectors –Platters –Heads –Cylinders Performance characteristics are: –Rotational latency (delay) – time to rotate to the required sector (RPM) –Seek time – time to find the required track –Transfer rate – time to transfer the data

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Magnetic Disk Layout Platter Track Platters Sectors Tracks COPYRIGHT 1998 MORGAN KAUFMANN PUBLISHERS, INC. ALL RIGHTS RESERVED

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Zone Bit Recording Originally, all tracks had the same number of sectors –Each sector = 512 Bytes Inefficient! Limited by the density of the smallest tracks Outer tracks can hold more sectors (store more data) than inner tracks  called Zone Bit Recording (ZBR) The sequence of information recorded on each sector is –sector number –gap –data including error correction code bits –gap

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Disk Performance What is the average time to read / write a 512 byte sector given: –Avg seek time = 9 ms –Rotation = 7200 rpm –Transfer rate = 150 MB / sec (e.g. serial ATA) Avg access time = seek time + rotational delay + transfer time = 9 ms + (½ rev) / 7200 rpm bytes / (150 MB / sec) = (½ * 60) / / (150 * 2 20 ) = = sec = ms Suppose the rotation rate was 5400 rpm? Avg Access time = (½ * 60) / = = ms

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Disk Reliability vs Availability In processing, our main concern is performance (and cost) In I/O, our main concern is reliability (and cost) Reliability  is anything broken? Availability  is the system still available to the user, even if it is broken? RAID technology is designed to provide increased availability for potentially unreliable devices RAID – Redundant Array of Inexpensive Disks –Patterson / Katz / Gibson

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID Provide a set of physical disks that appear to be a single logical drive Distribute the data across the drives in the array Allow levels of redundancy to permit recovery of failed disks 7 (basic) levels of RAID (RAID0 – RAID6) RAID is NOT a backup system! –Its made to maintain uptime through failures

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID0 No redundancy (therefore not really RAID at all) Allocate sectors (or larger units) from a single logical drive in stripes across multiple physical drives PhysicalDrive Physical Drive Physical Drive Physical Drive LogicalDrive stripes

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID0 Summary Called striping the data –Each stripe is  1 sector Advantage: –access to large sections of contiguous data can be done in parallel over all disks Disadvantage: –no redundancy

Physical Drive Physical Drive PhysicalDrive Physical Drive RAID1 Mirrored data Make a complete mirror (i.e. duplicate) of all data PhysicalDrive Physical Drive Physical Drive Physical Drive mirrors

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID1 Summary Uses stripes (of sectors) across drives just like RAID0 Replicates each disk with an exact copy Advantage: –Allows parallel access (just like RAID0) –Failure of any number of drives does not impact availability Disadvantage: –Writes must occur over both “arrays” –Very expensive (2 x disks)

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID2 Redundancy through Hamming codes Store enough extra data in order to detect and correct errors, or in order to provide availability in the case of a failed drive Stripes are small  1 bit (!) per stripe originally  later bytes/words All disk heads are synchronized – does not permit parallel access as in RAID0 and RAID1 Requires  log 2 (#disks) extra disks to implement

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Hamming Codes A technique for enabling error detection & correction (and therefore redundancy as well) Ex: ECC Scheme: –store 1110, but suppose an error changes this to Even parity bit Wrong parity bit! Error in data Error detected

Physical Drive 5 0xxx xxxx Physical Drive 4 1xxx xxxx Physical Drive 6 0xxx xxxx RAID2 Disk Layout Physical Drive 0 1xxx xxxx Physical Drive 1 1xxx xxxx Physical Drive 2 1xxx xxxx Physical Drive 3 0xxx xxxx Parity bits Data Stripes of bytes or words

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID2 Summary Error correction is done across disks Advantage: –Useful in a high failure environment Disadvantage: –Expensive –Modern disks do not exhibit high failure rate Not used in practice

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Physical Drive 4 1xxx xxxx RAID3 Bit interleaved parity Like RAID2 but use only 1 parity drive Physical Drive 0 1xxx xxxx Physical Drive 1 1xxx xxxx Physical Drive 2 1xxx xxxx Physical Drive 3 0xxx xxxx Stripes of bytes or words Parity drive

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID3 Summary Small stripes (byte or word) interleaved across disks Parity = XOR of all bits in the byte / word Advantage: –All disks participate in all data accesses  high transfer rate –If a single drive fails, lost bit (stripe) can be reconstructed easily Disadvantage: –Only 1 access per “time” –Can detect / correct only single failures

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Physical Drive 4 P(0-3) P(4-7) P(8-11) RAID4 Block level parity Same parity scheme as RAID3 but uses large blocks per stripe Physical Drive 0 Block0 Block4 Block8 Physical Drive 1 Block1 Block5 Block9 Physical Drive 2 Block2 Block6 Block10 Physical Drive 3 Block3 Block7 Block11 Stripes of large “blocks” Parity drive

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID4 Summary Interleaved blocks across disks Advantage: –Allows independent access due to large stripes  can support multiple independent reads –Error detection / correction supported up to single bit errors Disadvantage: –On parallel writes  penalty incurred since all writes require access to the same parity disk to update the parity Not used in practice

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Physical Drive 4 P(0-3) Block7 Block11 RAID5 Block level distributed parity Same scheme as RAID4 but parity is interleaved with the data Physical Drive 0 Block0 Block4 Block8 Physical Drive 1 Block1 Block5 Block9 Physical Drive 2 Block2 Block6 P(8-11) Physical Drive 3 Block3 P(4-7) Block10

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID5 Summary Interleaved blocks across disks and interleaved with data Advantage: –Can support multiple independent reads as long as access is to independent disks –Can support multiple independent writes as long as write and parity disks are independent –Error detection / correction supported up to single bit errors –Reduces I/O bottleneck Most common approach used in practice

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID6 Summary AKA “P + Q Redundancy” Same as RAID5 (interleaved blocks and interleaved parity) –But has a second set of parity bits Advantage: –Same as RAID5 –Plus the ability to tolerate a second failure (either disk or operator) Disadvantage: –Takes one more extra drive worth of space (more expensive) In case you want to be extra careful

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst RAID Systems – Failure and Recovery RAID1 and RAID5 are the most common in practice –RAID1  read performance/redundancy –RAID5  cost/redundancy (performance not bad either) If a RAID system loses a disk, how do you fix it? –When a disk fails, we’ve entered a “danger zone” where one more error causes data loss Repairs need to be made quickly to minimize this time –Hot-swappable drives are doable, but expensive/difficult –Some systems use standby spares which are automatically used as a replacement when an error occurs Once a new disk is provided, the RAID system will reconstruct the data from redundancy or parity information –All done during uptime

CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst What RAID Can and Cannot Do RAID Can protect uptime –Downtime = $$$ RAID Can improve performance on certain applications –Anytime you’re accessing/moving around large files, striping gives you lots of access bandwidth RAID Cannot protect the data on the array! –One file system  single point of failure –Corruption, Virus, Fire, Flood, etc. – Not an alternative to a separate backup! RAID Cannot improve performance on all applications