"1"1 Introduction to Managing Data " Describe problems associated with managing large numbers of disks " List requirements for easily managing large amounts.

Slides:



Advertisements
Similar presentations
By Rakshith Venkatesh Outline What is RAID? RAID configurations used. Performance of each configuration. Implementations. Way.
Advertisements

Redundant Array of Independent Disks (RAID) Striping of data across multiple media for expansion, performance and reliability.
A CASE FOR REDUNDANT ARRAYS OF INEXPENSIVE DISKS (RAID) D. A. Patterson, G. A. Gibson, R. H. Katz University of California, Berkeley.
Copyright © 2009 EMC Corporation. Do not Copy - All Rights Reserved.
What is RAID Redundant Array of Independent Disks.
1 Jason Drown Mark Rodden (Redundant Array of Inexpensive Disks) RAID.
RAID (Redundant Arrays of Independent Disks). Disk organization technique that manages a large number of disks, providing a view of a single disk of High.
RAID A RRAYS Redundant Array of Inexpensive Discs.
RAID Oh yes Whats RAID? Redundant Array (of) Independent Disks. A scheme involving multiple disks which replicates data across multiple drives. Methods.
RAID: Redundant Array of Inexpensive Disks Supplemental Material not in book.
RAID Redundant Array of Independent Disks
Module – 3 Data protection – raid
“Redundant Array of Inexpensive Disks”. CONTENTS Storage devices. Optical drives. Floppy disk. Hard disk. Components of Hard disks. RAID technology. Levels.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
© 2009 EMC Corporation. All rights reserved. EMC Proven Professional The #1 Certification Program in the information storage and management industry Data.
1 Magnetic Disks 1956: IBM (RAMAC) first disk drive 5 Mb – Mb/in $/year 9 Kb/sec 1980: SEAGATE first 5.25’’ disk drive 5 Mb – 1.96 Mb/in2 625.
R.A.I.D. Copyright © 2005 by James Hug Redundant Array of Independent (or Inexpensive) Disks.
Chapter 3 Presented by: Anupam Mittal.  Data protection: Concept of RAID and its Components Data Protection: RAID - 2.
WHAT IS RAID? Christopher J Dutra Seton Hall University.
Chapter 5: Server Hardware and Availability. Hardware Reliability and LAN The more reliable a component, the more expensive it is. Server hardware is.
Solaris Volume Manager M. Desouky. RAID Overview SDS Software SDS Installation SDS User Interfaces MD State Database Concats & Stripes Mirrors Hot Spares.
RAID Technology. Use Arrays of Small Disks? 14” 10”5.25”3.5” Disk Array: 1 disk design Conventional: 4 disk designs Low End High End Katz and Patterson.
REDUNDANT ARRAY OF INEXPENSIVE DISCS RAID. What is RAID ? RAID is an acronym for Redundant Array of Independent Drives (or Disks), also known as Redundant.
Computer ArchitectureFall 2007 © November 28, 2007 Karem A. Sakallah Lecture 24 Disk IO and RAID CS : Computer Architecture.
I/O Systems and Storage Systems May 22, 2000 Instructor: Gary Kimura.
CSE 451: Operating Systems Winter 2010 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura.
© 2009 IBM Corporation Statements of IBM future plans and directions are provided for information purposes only. Plans and direction are subject to change.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
By : Nabeel Ahmed Superior University Grw Campus.
Redundant Array of Inexpensive Disks (RAID). Redundant Arrays of Disks Files are "striped" across multiple spindles Redundancy yields high data availability.
ICOM 6005 – Database Management Systems Design Dr. Manuel Rodríguez-Martínez Electrical and Computer Engineering Department Lecture 6 – RAID ©Manuel Rodriguez.
RAID Ref: Stallings. Introduction The rate in improvement in secondary storage performance has been considerably less than the rate for processors and.
LAN / WAN Business Proposal. What is a LAN or WAN? A LAN is a Local Area Network it usually connects all computers in one building or several building.
Lecture 4 1 Reliability vs Availability Reliability: Is anything broken? Availability: Is the system still available to the user?
CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Storage Systems.
Storage and NT File System INFO333 – Lecture Mariusz Nowostawski Noria Foukia.
CSE 321b Computer Organization (2) تنظيم الحاسب (2) 3 rd year, Computer Engineering Winter 2015 Lecture #4 Dr. Hazem Ibrahim Shehata Dept. of Computer.
Redundant Array of Independent Disks
Two or more disks Capacity is the same as the total capacity of the drives in the array No fault tolerance-risk of data loss is proportional to the number.
1 Chapter 7: Storage Systems Introduction Magnetic disks Buses RAID: Redundant Arrays of Inexpensive Disks.
Lecture 9 of Advanced Databases Storage and File Structure (Part II) Instructor: Mr.Ahmed Al Astal.
Redundant Array of Inexpensive Disks aka Redundant Array of Independent Disks (RAID) Modified from CCT slides.
CSI-09 COMMUNICATION TECHNOLOGY FAULT TOLERANCE AUTHOR: V.V. SUBRAHMANYAM.
Guide to Linux Installation and Administration, 2e 1 Chapter 9 Preparing for Emergencies.
Module 9: Configuring Storage
4.1 © 2004 Pearson Education, Inc. Exam Managing and Maintaining a Microsoft® Windows® Server 2003 Environment Lesson 4: Organizing a Disk for Data.
Redundant Array of Independent Disks.  Many systems today need to store many terabytes of data.  Don’t want to use single, large disk  too expensive.
The concept of RAID in Databases By Junaid Ali Siddiqui.
Αρχιτεκτονική Υπολογιστών Ενότητα # 6: RAID Διδάσκων: Γεώργιος Κ. Πολύζος Τμήμα: Πληροφορικής.
RAID Systems Ver.2.0 Jan 09, 2005 Syam. RAID Primer Redundant Array of Inexpensive Disks random, real-time, redundant, array, assembly, interconnected,
WINDOWS SERVER 2003 Genetic Computer School Lesson 12 Fault Tolerance.
1 CEG 2400 Fall 2012 Network Servers. 2 Network Servers Critical Network servers – Contain redundant components Power supplies Fans Memory CPU Hard Drives.
Hands-On Microsoft Windows Server 2008 Chapter 7 Configuring and Managing Data Storage.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
RAID Technology By: Adarsha A,S 1BY08A03. Overview What is RAID Technology? What is RAID Technology? History of RAID History of RAID Techniques/Methods.
RAID TECHNOLOGY RASHMI ACHARYA CSE(A) RG NO
Network-Attached Storage. Network-attached storage devices Attached to a local area network, generally an Ethernet-based network environment.
© 2009 EMC Corporation. All rights reserved. Why do we need RAID o Performance limitation of disk drive o An individual drive has a certain life expectancy.
Fujitsu Training Documentation RAID Groups and Volumes
Operating System I/O System Monday, August 11, 2008.
Storage Virtualization
RAID RAID Mukesh N Tekwani
ICOM 6005 – Database Management Systems Design
CSE 451: Operating Systems Winter 2009 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura 1.
TECHNICAL SEMINAR PRESENTATION
UNIT IV RAID.
Mark Zbikowski and Gary Kimura
CSE 451: Operating Systems Winter 2012 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura 1.
RAID RAID Mukesh N Tekwani April 23, 2019
Seminar on Enterprise Software
Presentation transcript:

"1"1 Introduction to Managing Data " Describe problems associated with managing large numbers of disks " List requirements for easily managing large amounts of data " List techniques that facilitate managing large amounts of data " Define and differentiate high availability and fault tolerance " Describe commonly implemented RAID (redundant array of independent disks) levels

"2"2 Managing Large Numbers of Disks Servers today are configured with more disks than ever. This poses significant problems for system administrators: " Greater probability of disk failure: As the number of disks increases, so does the probability that one of them will fail. (MTBF (mean time between failures) is reduced.) " Partitioning of file systems: File systems are limited to the size of a single disk; to use a large number of disks, the file systems must partitioned so that at least one file system resides on each disk. " Longer reboot times: A busy system with many active disks can take an an unacceptably long time to run fsck after a system crash.

"3"3 Managing Large Sets of Data " Lower the number of file systems that must be managed: If file system can larger than one physical disk, it will not be necessary to artificially decompose a logical file system into multiple smaller file systems that each fit on a single disk. " Prevent failed disks from making data unavailable: The probability of a disk failure increases with the number of disks on a system. " More efficient balance the I/O load across the disks: Balancing the I/O performance. " Remove the need of check file systems at boot time: Performing file system checks (fsck) at boot time is time consuming.

"4"4 Managing Large Sets of Data " Allow file systems to grow while they are in use: Allowing file system to grow while they are in use reduces the system down time and eases the system administration burden (no need for backup/restore cycle. " Ease administration by providing a GUI to mask the underlying complexity " Allow dual-host failover configurations with redundant disks: In a dual-host failover configuration, one host can "take over" disk management for another failed host.

"5"5 Techniques for Managing Data

"6"6 " Integrated graphical user interface (GUI): More intuitive and easier to use " Concatenation: Two or more physical devices are combined into a single logical device " Expanding file Systems: Increasing the size of a UNIX file system while it is mounted and without disrupting access to the data " Hot Spares: A component set up to be automatically substituted for a failed component of a mirrored or RAID device. " Disk Striping: Data is interlaced among multiple physical devices; improves I/O performance by balancing the load. " RAID 5: Data and parity is interlaced among multiple physical devices. " Disk Mirroring: Multiple copies of the data are maintained on different physical devices.

"7"7 High Availability and Fault Tolerance " High availability: A system is said to have "high availability" (HA) if it can provide access to data most of the time while managing the integrity of that data " Fault tolerant: A system is said to be "fault tolerant" (FT) if it provides data integrity and continuos data availability. " High availability versus fault tolerant: Highly available and fault tolerant systems are separated by issues of function, cost, and design.

"8"8 Common RAID Implementations Some of the RAID levels are: " RAID 0:striping/concatenation " RAID 1:mirroring " RAID 0+1:striping plus mirroring " RAID 3:Striping with dedicated parity " RAID 5:Striping with distributed parity

"9"9 Concatenation -- RAID level 0 " Combines multiple physical disks into a single virtual disk " Address space is contiguous " No data redundancy

" 1010 Concatenation Summary " Write performance is the same; read performance may be improved if the reads are random. " One hundred percent of the disk capacity is available for user data. " There is no redundancy. " Concatenation is less reliable, as the loss of one disk ultimately results in the loss of data on all disks.

" 1111 Striping -- RAID level 0 " Data stream placed across multiple disks in equal-sized chunks " Improves I/O per second (IOPS) performance " Degrades reliability

" 1212 Striping Summary " Performance is improved; chunk sizes can be optimized for sequential or random access. " One hundred percent of the disk capacity is available for user data. " There is no redundancy. " Striping is less reliable, as the loss of one disk results in the loss of data on all striped disks.

" 1313 Mirroring -- RAID Level 1 " Fully redundant copy of the data on one or more disks (double the cost per megabyte of disk space) " All writes duplicated implies slower write performance " Both drives can be used for reads to improve performance

" 1414 Mirroring Summary " Performance may be improved on read performance, but will suffer on write performance. " The cost of a mirrored implementation is much higher than a standard disk system -- mirroring requires double the storage costs. " In the event of failure, applications can continue to use the remaining half of the mirror at close to full performance. " Recovering from a disk failure consists of simply duplicating the contents of the failed disk's mirror to a new drive.

" 1515 Striping and Mirroring--Raid level 0+1 " By combining mirroring and striping both high reliability and performance are provided (but at a high cost)

" 1616 Striped, Then Mirrored (RAID 0+1) " RAID 0+1 systems have both the improved performance of striping and improved reliability of mirroring. " RAID 0+1 systems suffer the high cost of mirrored systems, requiring twice the disk space of fully independent spindles. " RAID 0+1 systems can tolerate the failure of any single disk and continue to deliver data with virtually no performance degradation.

" 1717 Striping With Dedicated Parity--RAID Level 3 " Data is striped across a group of disks " One disk per group is dedicated to parity " Parity disk protects against any one disk of the group failing " All disks are read and written to simultaneously

" 1818 Striping With Dedicated Parity (Cont.,) Other Features " Data is striped across all spindles " Dedicated parity disk contains XOR (exclusive OR) of data disks " Bandwidth is equal to n-1 disk transfer rate " All actuators move in concert as the spindles are synchronized. " Single-chunk random I/O slows the RAID group to the performance of a single disk. XOR is commutative and associated across the equation: Cp = C1 (XOR) C2 (XOR) C3 and C2 = C1 (XOR) C3 (XOR) Cp, etc. Where do chunks 7, 8, and 9 go on the diagram? Where does the parity go? " RAID 3 provides good sequential transfer rates, but at the expense of random I/O performance. " RAID 3 requires only one additional drive beyond those used for data.

" 1919 Striping With Dedicated Parity (Cont.,) " If the parity drive fails, operations continue with no loss of performance (but there is no redundancy). If a data disk fails, the data is still available, but it must be calculated from the remaining data disks and the parity disk. " Recovery involves reading data from surviving disks, computing the exclusive OR, and writing the result to the replacement drive.

" 2020 Striping With Distributed Parity -- RAID level 5 " Both parity and data are striped across a group of drives " Each drive can be read independently " Parity protects against single disk failure

" 2121 RAID Level 5 Summary RAID 5 implements data protection through a distributed parity scheme. Additional features include: " Independent access is available to individual drives " Data and parity are both striped across spindles " Reads per second can reach disk rate times number of disks " Single-chunk writes require four disk operations: read old data, read old parity, calculate new parity, write new data, write new parity.

" 2222 RAID Level 5 Summary " Overall random I/O performance is dependent on percentage of writes. " RAID 5 requires only one additional drive beyond those used for data " Data can be accessed with a failed drive, with some performance penalties: A To read data from a surviving drive--No change. A To read data from a failed drive--Corresponding chunks from surviving drives in the stripe are read and linked together with Xor to derive the data. A To write to a surviving drive--If the failed drive holds parity data, the parity data, the write proceeds normally without calculating parity. If the failed drive holds data, then a read-modify-write sequence is required A To write to a failed drive--All the data from the surviving data drives are linked with the new data using XOR, and the result is written to the parity drive.

" 2323 RAID Level 5 Summary (contd.,) " Recovery requires that the data from the remaining chunks in the stripe be read, linked together with XOR, and the result written to the replacement drive.