Disk Scrubbing in Large Archival Storage Systems Thomas Schwarz, S.J. 1,2 Qin Xin 1,3, Ethan Miller 1, Darrell Long 1, Andy Hospodor 1,2, Spencer Ng 3.

Slides:



Advertisements
Similar presentations
By Rakshith Venkatesh Outline What is RAID? RAID configurations used. Performance of each configuration. Implementations. Way.
Advertisements

Disk Arrays COEN 180. Large Storage Systems Collection of disks to store large amount of data. Performance advantage: Each drive can satisfy only so many.
Redundant Array of Independent Disks (RAID) Striping of data across multiple media for expansion, performance and reliability.
A CASE FOR REDUNDANT ARRAYS OF INEXPENSIVE DISKS (RAID) D. A. Patterson, G. A. Gibson, R. H. Katz University of California, Berkeley.
Lecture 4: A Case for RAID (Part 2) Prof. Shahram Ghandeharizadeh Computer Science Department University of Southern California.
I/O Management and Disk Scheduling Chapter 11. I/O Driver OS module which controls an I/O device hides the device specifics from the above layers in the.
CSE 451: Operating Systems Spring 2012 Module 20 Redundant Arrays of Inexpensive Disks (RAID) Ed Lazowska Allen Center 570 ©
Triple-Parity RAID and Beyond Hai Lu. RAID RAID, an acronym for redundant array of independent disks or also known as redundant array of inexpensive disks,
RAID Redundant Array of Independent Disks
 RAID stands for Redundant Array of Independent Disks  A system of arranging multiple disks for redundancy (or performance)  Term first coined in 1987.
0 秘 Type of NAND FLASH Discuss the Differences between Flash NAND Technologies: SLC :Single Level Chip MLC: Multi Level Chip TLC: Tri Level Chip Discuss:
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
RAID Redundant Arrays of Inexpensive Disks –Using lots of disk drives improves: Performance Reliability –Alternative: Specialized, high-performance hardware.
Chapter 3 Presented by: Anupam Mittal.  Data protection: Concept of RAID and its Components Data Protection: RAID - 2.
Lecture 36: Chapter 6 Today’s topic –RAID 1. RAID Redundant Array of Inexpensive (Independent) Disks –Use multiple smaller disks (c.f. one large disk)
Reliability of Disk Systems. Reliability So far, we looked at ways to improve the performance of disk systems. Next, we will look at ways to improve the.
Performance/Reliability of Disk Systems So far, we looked at ways to improve the performance of disk systems. Next, we will look at ways to improve the.
Availability in Global Peer-to-Peer Systems Qin (Chris) Xin, Ethan L. Miller Storage Systems Research Center University of California, Santa Cruz Thomas.
Computer ArchitectureFall 2007 © November 28, 2007 Karem A. Sakallah Lecture 24 Disk IO and RAID CS : Computer Architecture.
1 Storage (cont’d) Disk scheduling Reducing seek time (cont’d) Reducing rotational latency RAIDs.
Failure Trends in a Large Disk Drive Population Authors: Eduardo Pinheiro, Wolf- Dietrich Weber and Luiz Andr´e Barroso Presented by Vinuthna & Arjun.
Section Disk Failures Kevin Grant
Performance Evaluation of Peer-to-Peer Video Streaming Systems Wilson, W.F. Poon The Chinese University of Hong Kong.
Secondary Storage CSCI 444/544 Operating Systems Fall 2008.
I/O Systems and Storage Systems May 22, 2000 Instructor: Gary Kimura.
RAID Systems CS Introduction to Operating Systems.
CSE 451: Operating Systems Winter 2010 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura.
1 Product Reliability Chris Nabavi BSc SMIEEE © 2006 PCE Systems Ltd.
UC Santa Cruz Providing High Reliability in a Minimum Redundancy Archival Storage System Deepavali Bhagwat Kristal Pollack Darrell D. E. Long Ethan L.
Using Algebraic Signatures in Storage Applications Thomas Schwarz, S.J. Associate Professor, Santa Clara University Associate, SSRC UCSC Storage Systems.
Reliability Analysis of An Energy-Aware RAID System Shu Yin Xiao Qin Auburn University.
Chapter 6 RAID. Chapter 6 — Storage and Other I/O Topics — 2 RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f.
RAID Shuli Han COSC 573 Presentation.
Redundant Array of Independent Disks
Introduction to Computer Networks Introduction to Computer Networks.
I/O – Chapter 8 Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
PARAID: The Gear-Shifting Power-Aware RAID Charles Weddle, Mathew Oldham, An-I Andy Wang – Florida State University Peter Reiher – University of California,
Redundant Array of Inexpensive Disks aka Redundant Array of Independent Disks (RAID) Modified from CCT slides.
Redundant Array of Independent Disks.  Many systems today need to store many terabytes of data.  Don’t want to use single, large disk  too expensive.
"1"1 Introduction to Managing Data " Describe problems associated with managing large numbers of disks " List requirements for easily managing large amounts.
Three-Dimensional Redundancy Codes for Archival Storage J.-F. Pâris, U. of Houston D. D. E. Long, U. C. Santa Cruz W. Litwin, U. Paris-Dauphine.
RAID Disk Arrays Hank Levy. 212/5/2015 Basic Problems Disks are improving, but much less fast than CPUs We can use multiple disks for improving performance.
Thomas Schwarz, S.J. Qin Xin, Ethan Miller, Darrell Long, Andy Hospodor, Spencer Ng Summarized by Leonid Kibrik.
Disk Failures Eli Alshan. Agenda Articles survey – Failure Trends in a Large Disk Drive Population – Article review – Conclusions – Criticism – Disk failure.
UC Santa Cruz Impact of Failure on Interconnection Networks for Large Storage Systems Qin Xin, Ethan L. Miller, Thomas J. E. Schwarz*, Darrell D. E. Long.
Hands-On Microsoft Windows Server 2008 Chapter 7 Configuring and Managing Data Storage.
Disk Failures Skip. Index 13.4 Disk Failures Intermittent Failures Organizing Data by Cylinders Stable Storage Error- Handling.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
Reliability of Disk Systems. Reliability So far, we looked at ways to improve the performance of disk systems. Next, we will look at ways to improve the.
I/O Errors 1 Computer Organization II © McQuain RAID Redundant Array of Inexpensive (Independent) Disks – Use multiple smaller disks (c.f.
Magnetic Disks Have cylinders, sectors platters, tracks, heads virtual and real disk blocks (x cylinders, y heads, z sectors per track) Relatively slow,
CS Introduction to Operating Systems
Storage HDD, SSD and RAID.
Hardware & Software Reliability
RAID Non-Redundant (RAID Level 0) has the lowest cost of any RAID
RAID Disk Arrays Hank Levy 1.
RAID RAID Mukesh N Tekwani
RAID Disk Arrays Hank Levy 1.
CSE 451: Operating Systems Spring 2005 Module 17 Redundant Arrays of Inexpensive Disks (RAID) Ed Lazowska Allen Center 570.
CSE 451: Operating Systems Autumn 2010 Module 19 Redundant Arrays of Inexpensive Disks (RAID) Ed Lazowska Allen Center 570.
CSE 451: Operating Systems Winter 2009 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura 1.
RAID Redundant Array of Inexpensive (Independent) Disks
Mark Zbikowski and Gary Kimura
CSE 451: Operating Systems Autumn 2004 Redundant Arrays of Inexpensive Disks (RAID) Hank Levy 1.
CSE 451: Operating Systems Winter 2012 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura 1.
CSE 451: Operating Systems Autumn 2009 Module 19 Redundant Arrays of Inexpensive Disks (RAID) Ed Lazowska Allen Center 570.
RAID Disk Arrays Hank Levy 1.
RAID RAID Mukesh N Tekwani April 23, 2019
CSE 451: Operating Systems Winter 2004 Module 17 Redundant Arrays of Inexpensive Disks (RAID) Ed Lazowska Allen Center 570.
CSE 451: Operating Systems Winter 2006 Module 18 Redundant Arrays of Inexpensive Disks (RAID) Ed Lazowska Allen Center 570.
Presentation transcript:

Disk Scrubbing in Large Archival Storage Systems Thomas Schwarz, S.J. 1,2 Qin Xin 1,3, Ethan Miller 1, Darrell Long 1, Andy Hospodor 1,2, Spencer Ng 3 1 Storage Systems Resource Center, U. of California, Santa Cruz 2 Santa Clara University, Santa Clara, CA 3 Hitachi Global Storage Technologies, San Jose Research Center,

Introduction Large archival storage systems: Protect data more proactively Keep disks powered off for long periods of time Have low rate of data access Protect data by storing it redundantly.

Introduction Failures can happen At the block level. At the device level. Failures may remain undetected for long periods of time. A failure may unmask one or more additional failures. Reconstruction procedure accesses data on other devices. Those devices can have suffered previous failures.

Introduction We investigate the efficacy of disk scrubbing. Disk Scrubbing accesses a disk to see whether the data can still be read. Reading a single block shows that the device still works. Reading all blocks shows that we can read all the data on the block.

Contents 1. Disk Failure Taxonomy 2. System Overview 3. Disk Scrubbing Modeling 4. Power Cycles and Reliability 5. Optimal Scrubbing Interval 6. Simulation Results

Disk Failure Taxonomy Disk Blocks 512B sector uses error control coding Read to a block successfully either corrects all errors, or retries and then: flags block as unreadable, or misreads block. Disk Failure Rates Depend highly on Environment: Temperature, Vibrations, Air quality Age. Vintage.

Disk Failure Taxonomy Block Failure Rate estimate: Since: 1/3 of all field returns for server drives are due to hard errors. RAID users (90%) do not return drives with hard errors. 10% of all disks sold account for 1/3 of all errors. Hence: Mean Time between Block Failures is 3/10 MTBF of all disk failures. Mean time to disk failure is 3/2 of MTBF. 1 million hour rated drive has 3*10 5 mean time between block failure. 1.5*10 6 mean time between disk failure. This is one back of the envelope calculation based on numbers by one anonymous disk manufacturer. The results seem to be accepted by many.

System Overview Disks are powered down when not in use. Use m+k redundancy scheme: Store data in large blocks. m blocks grouped into an r-group. Add k parity data blocks to r-group. Small blocks lead to fast reconstruction and good reconstruction load distribution. Large blocks have slightly better reliability.

System Overview Disk Scrubbing Scrub an S - block Can read one block  device not failed. Can read all blocks  can access all data. Can read and verify all blocks  data can be read correctly. Use “algebraic signatures” for that. Can even verify that parity data accurately reflects client data.

System Overview If a bad block is detected, we usually can reconstruct its contents with parity / mirrored data. Scrubbing finds the error before it can hurt you.

Modeling Scrubbing Random Scrubbing: Scrub an S-block at random. (Exponential distribution). Deterministic Scrubbing: Scrub an S-block at regular intervals.

Modeling Scrubbing Opportunistic Scrubbing: Try to scrub when you access the disk anyway. “Piggyback scrubs on disk accesses” Efficiency depends on the frequency of accesses. MTBA: Mean Time Between Accesses (10 3 hours). Average scrub interval 10 4 hours. Block MTBF 10 5 hours.

Power Cycling and Reliability Turning a disk on or off has a significant impact. Even if disks move actuators away from surface (laptop disks). No direct data to measure impact of Power On Hours (POH). Extrapolate from Seagate data: One on / off cycle is roughly equivalent to running a disk for eight hours.

Determining Scrubbing Intervals Interval too short: Too much traffic  Disks busy  Increased error rate  Lower system MTBF. Interval too long: A failure more likely to unmask other failures.  More failures catastrophic.  Lower system MTBF.

Determining Scrubbing Intervals Mirrored reliability block N = 250 disks. Device failure rate: 5·10 5 hours Block failure rate: Time to read disk: 4 hours. Deterministic: without considering power-up effects. Deterministic with cycling: considering power-up effects. Opportunistic does not pay power-on penalty, but runs disk longer. Random does not pay power-on penalty. Random with cycling would be below the deterministic with cycling graph.

Determining Scrubbing Intervals Scrub frequently: You never know what you might find. Mirrored disks using opportunistic scrubbing (no power-on penalty). Assumes a high disk access rate.

Simulation Results 1PB archival data store. Disks have MTBF of 10 5 hours. 10,000 disk drives 10GB reliability blocks. ~1TB/day traffic

Simulation Results Two-way Mirroring

Simulation Results RAID 5 redundancy scheme

Simulation Results Mirroring. Opportunistic scrubbing with ~ three disk accesses per year. Observe that additional scrubbing leads to more power-on cycles that slightly increase occurrence of data losses.

Conclusions We have shown that disk scrubbing is a necessity for very large scale storage systems. Our simulations show the impact of power-on / power-off on reliability. We also note that lack of numbers on disk drive reliability prevents public research.