XtremIO Data Protection (XDP) Explained

Slides:



Advertisements
Similar presentations
Storage Management Lecture 7.
Advertisements

Redundant Array of Independent Disks (RAID) Striping of data across multiple media for expansion, performance and reliability.
DISK FAILURES PROF. T.Y.LIN CS-257 Presenter: Shailesh Benake(104)
Copyright © 2009 EMC Corporation. Do not Copy - All Rights Reserved.
What is RAID Redundant Array of Independent Disks.
1 Jason Drown Mark Rodden (Redundant Array of Inexpensive Disks) RAID.
RAID (Redundant Arrays of Independent Disks). Disk organization technique that manages a large number of disks, providing a view of a single disk of High.
Module – 3 Data protection – raid
CSCE430/830 Computer Architecture
Lecture 12 Reduce Miss Penalty and Hit Time
 RAID stands for Redundant Array of Independent Disks  A system of arranging multiple disks for redundancy (or performance)  Term first coined in 1987.
0 秘 Type of NAND FLASH Discuss the Differences between Flash NAND Technologies: SLC :Single Level Chip MLC: Multi Level Chip TLC: Tri Level Chip Discuss:
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
© 2009 EMC Corporation. All rights reserved. EMC Proven Professional The #1 Certification Program in the information storage and management industry Data.
RAID Redundant Arrays of Inexpensive Disks –Using lots of disk drives improves: Performance Reliability –Alternative: Specialized, high-performance hardware.
R.A.I.D. Copyright © 2005 by James Hug Redundant Array of Independent (or Inexpensive) Disks.
Chapter 3 Presented by: Anupam Mittal.  Data protection: Concept of RAID and its Components Data Protection: RAID - 2.
2P13 Week 11. A+ Guide to Managing and Maintaining your PC, 6e2 RAID Controllers Redundant Array of Independent (or Inexpensive) Disks Level 0 -- Striped.
CSE521: Introduction to Computer Architecture Mazin Yousif I/O Subsystem RAID (Redundant Array of Independent Disks)
Sean Traber CS-147 Fall  7.9 RAID  RAID Level 0  RAID Level 1  RAID Level 2  RAID Level 3  RAID Level 4 
REDUNDANT ARRAY OF INEXPENSIVE DISCS RAID. What is RAID ? RAID is an acronym for Redundant Array of Independent Drives (or Disks), also known as Redundant.
Reliability of Disk Systems. Reliability So far, we looked at ways to improve the performance of disk systems. Next, we will look at ways to improve the.
CPSC-608 Database Systems Fall 2008 Instructor: Jianer Chen Office: HRBB 309B Phone: Notes #6.
1 CS143: Disks and Files. 2 System Architecture CPU Main Memory Disk Controller... Disk Word (1B – 64B) ~ x GB/sec Block (512B – 50KB) ~ x MB/sec System.
Cse Feb-001 CSE 451 Section February 24, 2000 Project 3 – VM.
CSE 451: Operating Systems Winter 2010 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura.
“Better together” PowerVault virtualization solutions
By : Nabeel Ahmed Superior University Grw Campus.
Storage System: RAID Questions answered in this lecture: What is RAID? How does one trade-off between: performance, capacity, and reliability? What is.
Redundant Array of Inexpensive Disks (RAID). Redundant Arrays of Disks Files are "striped" across multiple spindles Redundancy yields high data availability.
ICOM 6005 – Database Management Systems Design Dr. Manuel Rodríguez-Martínez Electrical and Computer Engineering Department Lecture 6 – RAID ©Manuel Rodriguez.
DISKS IS421. DISK  A disk consists of Read/write head, and arm  A platter is divided into Tracks and sector  The R/W heads can R/W at the same time.
Lecture 4 1 Reliability vs Availability Reliability: Is anything broken? Availability: Is the system still available to the user?
1 Recitation 8 Disk & File System. 2 Disk Scheduling Disks are at least four orders of magnitude slower than main memory –The performance of disk I/O.
Redundant Array of Independent Disks
Two or more disks Capacity is the same as the total capacity of the drives in the array No fault tolerance-risk of data loss is proportional to the number.
N-Tier Client/Server Architectures Chapter 4 Server - RAID Copyright 2002, Dr. Ken Hoganson All rights reserved. OS Kernel Concept RAID – Redundant Array.
1 Chapter 7: Storage Systems Introduction Magnetic disks Buses RAID: Redundant Arrays of Inexpensive Disks.
Lecture 9 of Advanced Databases Storage and File Structure (Part II) Instructor: Mr.Ahmed Al Astal.
Parity Logging O vercoming the Small Write Problem in Redundant Disk Arrays Daniel Stodolsky Garth Gibson Mark Holland.
Redundant Array of Inexpensive Disks aka Redundant Array of Independent Disks (RAID) Modified from CCT slides.
Module – 4 Intelligent storage system
4.1 © 2004 Pearson Education, Inc. Exam Managing and Maintaining a Microsoft® Windows® Server 2003 Environment Lesson 4: Organizing a Disk for Data.
Mark A. Magumba Storage Management. What is storage An electronic place where computer may store data and instructions for retrieval The objective of.
1/14/2005Yan Huang - CSCI5330 Database Implementation – Storage and File Structure Storage and File Structure.
Multi-level Raid Multi-level Raid 2 Agenda Background -Definitions -What is it? -Why would anyone want it? Design Issues -Configuration and.
"1"1 Introduction to Managing Data " Describe problems associated with managing large numbers of disks " List requirements for easily managing large amounts.
+ CS 325: CS Hardware and Software Organization and Architecture Memory Organization.
The concept of RAID in Databases By Junaid Ali Siddiqui.
Αρχιτεκτονική Υπολογιστών Ενότητα # 6: RAID Διδάσκων: Γεώργιος Κ. Πολύζος Τμήμα: Πληροφορικής.
RAID Systems Ver.2.0 Jan 09, 2005 Syam. RAID Primer Redundant Array of Inexpensive Disks random, real-time, redundant, array, assembly, interconnected,
1 © 2002 hp Introduction to EVA Keith Parris Systems/Software Engineer HP Services Multivendor Systems Engineering Budapest, Hungary 23May 2003 Presentation.
CS399 New Beginnings Jonathan Walpole. Disk Technology & Secondary Storage Management.
Hands-On Microsoft Windows Server 2008 Chapter 7 Configuring and Managing Data Storage.
Part IV I/O System Chapter 12: Mass Storage Structure.
RAID Presentation Raid is an acronym for “Redundant array of independent Drives”, or Redundant array of inexpensive drives”. The main concept of RAID is.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
DATABASE OPERATORS AND SOLID STATE DRIVES Geetali Tyagi ( ) Mahima Malik ( ) Shrey Gupta ( ) Vedanshi Kataria ( )
Reliability of Disk Systems. Reliability So far, we looked at ways to improve the performance of disk systems. Next, we will look at ways to improve the.
Network-Attached Storage. Network-attached storage devices Attached to a local area network, generally an Ethernet-based network environment.
© 2009 EMC Corporation. All rights reserved. Why do we need RAID o Performance limitation of disk drive o An individual drive has a certain life expectancy.
RAID.
Fujitsu Training Documentation RAID Groups and Volumes
Vladimir Stojanovic & Nicholas Weaver
RAID RAID Mukesh N Tekwani
ICOM 6005 – Database Management Systems Design
TECHNICAL SEMINAR PRESENTATION
CS333 Intro to Operating Systems
RAID RAID Mukesh N Tekwani April 23, 2019
Storage Management Lecture 7.
Presentation transcript:

XtremIO Data Protection (XDP) Explained View this presentation in Slide Show mode

XDP Benefits Combines the best traits of traditional RAID with none of its drawbacks Ultra-low 8% fixed capacity overhead No RAID levels, stripe sizes, chunk sizes, etc. High levels of data protection Sustains up to two simultaneous failures per DAE* Multiple consecutive failures (with adequate free capacity) “Hot Space” - spare capacity is distributed (no hot spares) Rapid rebuild times Superior flash endurance Predictable, consistent, sub-millisecond performance *v2.2 encodes data for N+2 redundancy and supports a single rebuild per DAE. A future XIOS release will add double concurrent rebuild support.

XDP Stripe – Logical View 2 Parity columns C1 C2 C3 C4 C5 C6 C7 P Q 6 Data rows P1 Q1 P – is a column that contains parity per row The following slides show a simplified example of XDP. In reality, XDP uses a (23+2) x 28 stripe. Q – is a column that contains parity per diagonal. P2 Q2 P3 Q3 P4 Q4 P5 Q5 P6 Q6 XDP strip here is reduced in size, this is to simplify the explanations of the XDP technology Q7 4K 7 Data columns Every block in the XDP stripe is 4KB in size.

Each SSD contains the same numbers of P and Q columns Physical View Although each column is represented in this diagram as a logical block, The system has the ability to read or write in granularity of 4KB or less C1 C2 C3 C4 C5 C6 C7 P Q C Each SSD contains the same numbers of P and Q columns Stripe’s columns are randomly distributed across the SSDs to avoid hot spots and congestion

SSD Failure Remaining data blocks are recovered using the diagonal parity, blocks previously read and stored in the controller memory, along with minimal reads from SSD XDP always reads the first two rows in a stripe and recovers C1’s blocks using row parity stored at P If the SSD where C1 is stored has failed, let’s see how XDP efficiently recovers the stripe The data is recovered, using the Q parity and data blocks from C2 and C3 that are already in the Storage Controller memory XDP minimizes reads required to recover data by 25% (30 vs. 42) increasing rebuild performance compared with traditional RAID. The system reads the rest of the diagonal data (columns C5, C6 and C7), and computes the value of C1 Next, XDP recovers data using the diagonal parity Q. It first reads the parity information from row Q Expedited recovery process completes with fewer reads and parity compute cycles. C1 C2 C3 C4 C5 C6 C7 P Q Controller Memory C1 C2 C3 C4 C5 C6 C7 P P Q The C1 column will be placed at a spare capacity in the system. Note: this slide presents the XDP capabilities, not all of those capabilities will be available at first GA (but are planned for future release). Number of Writes Number of Reads 6 5 2 1 4 3 15 14 7 30 22 23 24 19 18 26 27

XDP Rebuilds & Hot Space Allows SSDs to fail-in-place Rapid rebuilds No performance impact after rebuild completes for up to five failed SSDs per X-Brick 3 failed SSDs ~330K IOPS 4 failed SSDs ~330K IOPS 5 failed SSDs ~330K IOPS

Stripe update at 80% utilization % Free Blocks Example shows new I/Os overwriting addresses with existing data – there is no net increase in capacity consumed (space frees up in other stripes) At least one stripe is guaranteed to be 40% empty => hosts benefit from the performance of a 40% empty array vs. a 20% empty array Re-ranking stripes according to % of free blocks Subsequent updates are performed using this algorithm Diagram shows an array that is 80% full Stripe number S9 S8 S7 S6 S5 S4 S3 S2 S1 40% 0% 0% The system ranks stripes according to utilization level Always writes to the stripe that is most free Writes to SSD as soon as enough blocks arrive to fill the entire emptiest stripe in the system (in this example 17 blocks are required) 40% Stripe Number % Free Blocks S3 0% S8 S2 S6 20% S5 S1 S9 40% S7 S4 Stripe Number % Free Blocks S9 0% S8 S2 S6 20% S5 S1 S7 40% S4 S3 20% 20% 40% 40% 0% 0% 20%

XDP Stripe - the Real Numbers Number of 4KB data blocks in a stripe Amount of data in a stripe Amount of Parity blocks in a stripe Total number of blocks in a stripe Total number of stripes in one X-Brick 28 X 23 = 644 4KB X 644= 2576KB 28 + 29 = 57 644 + 57 = 701 7.5TB per X-Brick/2,024KB ≈ 3M stripes 23 data columns Parity P Q RAID Overhead (of P,Q) = 57/701 = 8% The previous slides used an XDP stripe that is based on 7 x 6 data blocks, in reality the system uses 28 X 23 data blocks 28 data rows 25 SSDs

Update Overhead Compared RAID Scheme Reads per Update Writes per Update Capacity Overhead RAID-5 2 N + 1 RAID-6 3 N + 2 RAID-1 N × 2 XtremIO (at 80%) 1.22 XtremIO (at 90%) 1.44