CSE431 Chapter 6A.1Irwin, PSU, 2008 Chapter 6A: Disk Systems Mary Jane Irwin ( www.cse.psu.edu/~mji )www.cse.psu.edu/~mji [Adapted from Computer Organization.

Slides:



Advertisements
Similar presentations
Lecture Objectives: 1)Define the terms track, sector, and seek as pertaining to a nonvolatile storage medium. 2)Calculate the average rotational for a.
Advertisements

CPS216: Data-Intensive Computing Systems Data Access from Disks Shivnath Babu.
1 Lecture 22: I/O, Disk Systems Todays topics: I/O overview Disk basics RAID Reminder: Assignment 8 due Tue 11/21.
IT253: Computer Organization
CS61C L16 Disks © UC Regents 1 CS161 Lecture 25 - Disks Laxmi Bhuyan
Secondary Storage Devices: Magnetic Disks
I/O Chapter 8. Outline Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
CS224 Spring 2011 Computer Organization CS224 Chapter 6A: Disk Systems With thanks to M.J. Irwin, D. Patterson, and J. Hennessy for some lecture slide.
13.2 Disks Gaurav Sharma Class ID Mechanics of Disks 2 Moving Principal Moving pieces of Disk are: 1. Disk assembly & 2. Head Assembly The.
Section 6.2. Record data by magnetizing the binary code on the surface of a disk. Data area is reusable Allows for both sequential and direct access file.
Princess Sumaya Univ. Computer Engineering Dept. Chapter 6:
Lecture 21Comp. Arch. Fall 2006 Chapter 8: I/O Systems Adapted from Mary Jane Irwin at Penn State University for Computer Organization and Design, Patterson.
CS1104 – Computer Organization
CS61C L13 I/O © UC Regents 1 CS 161 Chapter 8 - I/O Lecture 17.
Storing Data: Disks and Files: Chapter 9
CPE 442 io.1 Introduction To Computer Architecture CpE 442 I/O Systems.
Storage. The Memory Hierarchy fastest, but small under a microsecond, random access, perhaps 2Gb Typically magnetic disks, magneto­ optical (erasable),
CS4432: Database Systems II Data Storage - Lecture 2 (Sections 13.1 – 13.3) Elke A. Rundensteiner.
13.2 Disks Mechanics of Disks Presented by Chao-Hsin Shih Feb 21, 2011.
Review of Mem. HierarchyCSCE430/830 Review of Memory Hierarchy CSCE430/830 Computer Architecture Lecturer: Prof. Hong Jiang Courtesy of Yifeng Zhu (U.
11/18/2004Comp 120 Fall November 3 classes to go No class on Tuesday 23 November Last 2 classes will be survey and exam review Interconnect and.
11/29/2005Comp 120 Fall November 4 to go! Questions? Interrupts and I/O devices.
SECTIONS 13.1 – 13.3 Sanuja Dabade & Eilbroun Benjamin CS 257 – Dr. TY Lin SECONDARY STORAGE MANAGEMENT.
CS4432: Database Systems II Lecture 2 Timothy Sutherland.
Disks.
S.1 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output Cache Main Memory Secondary Memory (Disk)
12/3/2004EE 42 fall 2004 lecture 391 Lecture #39: Magnetic memory storage Last lecture: –Dynamic Ram –E 2 memory This lecture: –Future memory technologies.
1 CS222: Principles of Database Management Fall 2010 Professor Chen Li Department of Computer Science University of California, Irvine Notes 01.
I/0 devices.
SECTIONS 13.1 – 13.3 Sanuja Dabade & Eilbroun Benjamin CS 257 – Dr. TY Lin SECONDARY STORAGE MANAGEMENT.
CPSC 231 Secondary storage (D.H.)1 Learning Objectives Understanding disk organization. Sectors, clusters and extents. Fragmentation. Disk access time.
Introduction to Database Systems 1 The Storage Hierarchy and Magnetic Disks Storage Technology: Topic 1.
Magnetic Disk Storage Optical Laser Discs Damian Brain.
CSE431 Chapter 6A.1Irwin, PSU, 2008 CSE 431 Computer Architecture Fall 2008 Chapter 6A: Disk Systems Mary Jane Irwin ( )
CSE431 Chapter 6A.1Irwin, PSU, 2008 Chapter 6A: Disk Systems Mary Jane Irwin ( ) [Adapted from Computer Organization.
Silberschatz, Galvin and Gagne ©2013 Operating System Concepts – 9 th Edition Chapter 10: Mass-Storage Systems.
1 6 Further System Fundamentals (HL) 6.2 Magnetic Disk Storage.
I/O – Chapter 8 Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
1 Secondary Storage Management Submitted by: Sathya Anandan(ID:123)
CS2100 Computer Organisation Input/Output (AY2010/2011) Semester 2 Adapted from.
2.1 Operating System Concepts Chapter 2: Computer-System Structures Computer System Operation Storage Structure Storage Hierarchy Hardware Protection General.
I/O Computer Organization II 1 Introduction I/O devices can be characterized by – Behavior: input, output, storage – Partner: human or machine – Data rate:
August 1, 2001Systems Architecture II1 Systems Architecture II (CS ) Lecture 9: I/O Devices and Communication Buses * Jeremy R. Johnson Wednesday,
© 2005 Pearson Addison-Wesley. All rights reserved Figure 2.1 This chapter focuses on key hardware layer components.
CS4432: Database Systems II Data Storage 1. Storage in DBMSs DBMSs manage large amounts of data How does a DBMS store and manage large amounts of data?
Csci 136 Computer Architecture II – IO and Storage Systems Xiuzhen Cheng
Disk Basics CS Introduction to Operating Systems.
CS 101 – Sept. 28 Main vs. secondary memory Examples of secondary storage –Disk (direct access) Various types Disk geometry –Flash memory (random access)
Operating Systems (CS 340 D) Princess Nora University Faculty of Computer & Information Systems Computer science Department.
Section 13.2 – Secondary storage management (Former Student’s Note)
Disk Average Seek Time. Multi-platter Disk platter Disk read/write arm read/write head.
Section 13.2 – Secondary storage management. Index 13.2 Disks Mechanics of Disks The Disk Controller Disk Access Characteristics.
Magnetic Disk Rotational latency Example Find the average rotational latency if the disk rotates at 20,000 rpm.
CPSC 231 Secondary storage (D.H.)1 Learning Objectives Understanding disk organization. Sectors, clusters and extents. Fragmentation. Disk access time.
COSC 6340: Disks 1 Disks and Files DBMS stores information on (“hard”) disks. This has major implications for DBMS design! » READ: transfer data from disk.
Mohamed Younis CMCS 411, Computer Architecture 1 CMCS Computer Architecture Lecture 25 I/O Systems May 2,
CMSC 611: Advanced Computer Architecture I/O & Storage Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
Data Storage and Querying in Various Storage Devices.
Chapter 10: Mass-Storage Systems
CHAPTER 10: Computer Peripherals
Section 13.2 – Secondary storage management (Former Student’s Note)
Chapter 2: Computer-System Structures
IT 251 Computer Organization and Architecture
Part V Memory System Design
Disks and Files DBMS stores information on (“hard”) disks.
Sanuja Dabade & Eilbroun Benjamin CS 257 – Dr. TY Lin
Operating Systems (CS 340 D)
Input-output I/O is very much architecture/system dependent
Chapter 11: Mass-Storage Systems
Presentation transcript:

CSE431 Chapter 6A.1Irwin, PSU, 2008 Chapter 6A: Disk Systems Mary Jane Irwin ( ) [Adapted from Computer Organization and Design, 4 th Edition, Patterson & Hennessy, © 2008, MK] Read Sections 6.1, 6.2, 6.3, 6.9

CSE431 Chapter 6A.2Irwin, PSU, 2008 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output Cache Main Memory Secondary Memory (Disk)

CSE431 Chapter 6A.3Irwin, PSU, 2008 Magnetic Disk Purpose l Long term, nonvolatile storage l Lowest level in the memory hierarchy -slow, large, inexpensive General structure l A rotating platter coated with a magnetic surface l A moveable read/write head to access the information on the disk Track Sector

CSE431 Chapter 6A.4Irwin, PSU, 2008 Magnetic Disk Typical numbers l 1 to 4 platters (each with 2 recordable surfaces) per disk of 1 to 3.5 in diameter l Rotational speeds of 5,400 to 15,000 RPM l 10,000 to 50,000 tracks per surface -cylinder - all the tracks under the head at a given point on all surfaces l 100 to 500 sectors per track -the smallest unit that can be read/written (typically 512B) Track Sector

CSE431 Chapter 6A.5Irwin, PSU, 2008 Magnetic Disk Characteristic Disk read/write components 1.Seek time: time to position the head over the proper track (3 to 13 ms avg) -due to locality of disk references the actual average seek time may be only 25% to 33% of the advertised number 2.Rotational latency: wait for the desired sector to rotate under the head (½ of 1/RPM converted to ms) -0.5/5400RPM = 5.6ms to 0.5/15000RPM = 2.0ms Sector Track Cylinder Head Platter Controller + Cache

CSE431 Chapter 6A.6Irwin, PSU, 2008 Magnetic Disk Characteristic (cont. I) Disk read/write components 3.Transfer time: transfer a block of bits (one or more sectors) under the head to the disk controllers cache (70 to 125 MB/s are typical disk transfer rates in 2008) -the disk controllers cache takes advantage of spatial locality in disk accesses –cache transfer rates are much faster (e.g., 375 MB/s) Sector Track Cylinder Head Platter Controller + Cache

CSE431 Chapter 6A.7Irwin, PSU, 2008 Magnetic Disk Characteristic (cont. II) Disk read/write components 4.Controller time: the overhead the disk controller imposes in performing a disk I/O access (typically <.2 ms) Sector Track Cylinder Head Platter Controller + Cache

CSE431 Chapter 6A.8Irwin, PSU, 2008 Typical Disk Access Time If the measured average seek time is 25% of the advertised average seek time, then Avg disk read/write = 4.0 ms + 0.5/(15,000RPM/(60sec/min))+ 0.5KB/(100MB/sec) ms = = 6.2 ms Avg disk read/write = = 3.2 ms Calculate the average time to read or write a 512B sector for a disk rotating at 15,000 RPM with average seek time of 4 ms, a 100MB/sec transfer rate, and a 0.2 ms controller overhead The rotational latency is usually the largest component of the access time

CSE431 Chapter 6A.9Irwin, PSU, 2008 Disk Interface Standards Higher-level disk interfaces have a microprocessor disk controller that can lead to performance optimizations l SCSI (Small Computer Systems Interface) – l A set of standards for physically connecting and transferring data between computers and peripheral devices. Most commonly used for hard disks and tape drives. -commands -Protocols -electrical and optical interfaces

CSE431 Chapter 6A.10Irwin, PSU, 2008 Disk Interface Standards (continued) In particular, disk controllers have SRAM disk caches which support fast access to data that was recently read and often also include prefetch algorithms to try to anticipate demand

CSE431 Chapter 6A.11Irwin, PSU, 2008 Magnetic Disk Examples ( FeatureSeagate ST NS Seagate ST973451SS Seagate ST AS Disk diameter (inches) Capacity (GB) # of surfaces (heads)422 Rotation speed (RPM)7,20015,0005,400 Transfer rate (MB/sec) Minimum seek (ms)0.8r-1.0w0.2r-0.4w1.5r-2.0w Average seek (ms)8.5r-9.5w2.9r-3.3w12.5r-13.0w MTTF o C)1,200,0001,600,000?? Dim (inches), Weight (lbs) 1x4x5.8, x2.8x3.9, x2.8x3.9, 0.2 GB/cu.inch, GB/watt43, 9111, 937, 84 Power: op/idle/sb (watts)11/8/18/5.8/-1.9/0.6/0.2 Price in 2008, $/GB~$0.3/GB~$5/GB~$0.6/GB

CSE431 Chapter 6A.12Irwin, PSU, 2008 Disk Latency & Bandwidth Milestones Disk latency is one average seek time plus the rotational latency. Disk bandwidth is the peak transfer time of formatted data from the media (not from the cache). CDC WrenSG ST41SG ST15SG ST39SG ST37 RSpeed (RPM) Year Capacity (Gbytes) Diameter (inches) InterfaceST-412SCSI Bandwidth (MB/s) Latency (msec) Patterson, CACM Vol 47, #10, 2004

CSE431 Chapter 6A.13Irwin, PSU, 2008 Latency & Bandwidth Improvements In the time that the disk bandwidth doubles the latency improves by a factor of only 1.2 to 1.4

CSE431 Chapter 6A.14Irwin, PSU, 2008 Dependability, Reliability, Availability (Section 6.2) Reliability – measured by the mean time to failure (MTTF). Service interruption is measured by mean time to repair (MTTR) Availability – a measure of service accomplishment Availability = MTTF/(MTTF + MTTR)

CSE431 Chapter 6A.15Irwin, PSU, 2008 Dependability, Reliability, Availability (continued) To increase MTTF, either improve the quality of the components or design the system to continue operating in the presence of faulty components 1. Fault avoidance: preventing fault occurrence by construction 2. Fault tolerance: using redundancy to correct or bypass faulty components (hardware) l Fault detection versus fault correction l Permanent faults versus transient faults

CSE431 Chapter 6A.16Irwin, PSU, 2008 RAIDs: Disk Arrays (Read Section 6.9) Arrays of small and inexpensive disks l Increase potential throughput by having many disk drives -Data is spread over multiple disk -Multiple accesses are made to several disks at a time Redundant Array of Inexpensive Disks

CSE431 Chapter 6A.17Irwin, PSU, 2008 RAIDs: Disk Arrays (continued) Reliability is lower than a single disk But availability can be improved by adding redundant disks (RAID) l Lost information can be reconstructed from redundant information l MTTR: mean time to repair is in the order of hours l MTTF: mean time to failure of disks is tens of years Redundant Array of Inexpensive Disks

CSE431 Chapter 6A.18Irwin, PSU, 2008 RAID: Level 0 (No Redundancy; Striping) Multiple smaller disks as opposed to one big disk l striping Allocation of logically sequential blocks to separate disks l multiple blocks can be accessed in parallel increasing the performance l Same cost as one big disk – assuming 2 small disks cost the same as one big disk No redundancy, so what if one disk fails? l Failure of two or more disks is more likely as the number of disks in the system increases

CSE431 Chapter 6A.19Irwin, PSU, 2008 RAID: Level 1 (Redundancy via Mirroring) Uses twice as many disks as RAID 0 Whenever data is written to one disk, that data is also written to a redundant disk. So there are always two copies of the data l # redundant disks = # of data disks so twice the cost of one big disk What if one disk fails? l If a disk fails, the system just goes to the mirror for the data redundant (check) data Block 1Block 3Block 2Block 4Block 1Block 2Block 3Block 4 19

CSE431 Chapter 6A.20Irwin, PSU, 2008 RAID: Level 3 (Bit-Interleaved Parity) Cost of higher availability is reduced to 1/N where N is the number of disks in a protection group l # redundant disks = 1 × # of protection groups -writes require writing the new data to the data disk as well as computing the parity, meaning reading the other disks, so that the parity disk can be updated Can tolerate limited (single) disk failure, since the data can be reconstructed -reads require reading all the operational data disks as well as the parity disk to calculate the missing data that was stored on the failed disk blk1_bitblk3_bitblk2_bitblk4_bit 1 disk fails (odd) bit parity disk 1001 protection group :A group of data disks that share a common check desk Reads or writes go to all disks in the group, with an extra disk for parity 20

CSE431 Chapter 6A.21Irwin, PSU, 2008 RAID: Level 4 (Block-Interleaved Parity) Cost of higher availability still only 1/N but the parity is stored as blocks associated with sets of data blocks Think of the parity information as the parity of blk1 data (1 bit) and bk2 data (1 bit) etc. Could then take the parity of these 4 bits and store 1 parity bit associated with the parity of the 4 bits of the blocks in the protection group. Block parity disk blk1blk2blk3blk4 21

CSE431 Chapter 6A.22Irwin, PSU, 2008 RAID: Level 4 (Block-Interleaved Parity) (continued) Four times the throughput (striping) # redundant disks = 1 × # of protection groups Supports small reads and small writes (reads and writes that go to just one (or a few) data disk in a protection group) l by watching which bits change when writing new information, need only to change the corresponding bits on the parity disk l the parity disk must be updated on every write, so it is a bottleneck for back-to-back writes Block parity disk blk1blk2blk3blk4 22

CSE431 Chapter 6A.23Irwin, PSU, 2008 RAID: Level 4 (Block-Interleaved Parity) (continued) Can tolerate limited (one) disk failure, since the data can be reconstructed Block parity disk blk1blk2blk3blk4 23

CSE431 Chapter 6A.24Irwin, PSU, 2008 Small Writes Naïve RAID 4 small writes New D1 data D1D2D3D4P D1D2D3D4P 3 reads and 2 writes involving all the disks Optimized RAID 4 small writes New D1 data D1D2D3D4P D1D2D3D4P 2 reads and 2 writes involving just two disks 24

CSE431 Chapter 6A.25Irwin, PSU, 2008 RAID: Level 5 (Distributed Block-Interleaved Parity) Cost of higher availability still only 1/N but the parity block can be located on any of the disks so there is no single bottleneck for writes l Still four times the throughput (striping) l # redundant disks = 1 × # of protection groups l Supports small reads and small writes l Allows multiple simultaneous writes as long as the accompanying parity blocks are not located on the same disk Can tolerate limited disk failure, since the data can be reconstructed one of these assigned as the block parity disk 25

CSE431 Chapter 6A.26Irwin, PSU, 2008 Distributing Parity Blocks By distributing parity blocks to all disks, some small writes can be performed in parallel P P P P3 RAID 4RAID P P P P Time Can be done in parallel 26