Graduate Computer Architecture I Lecture 13: Storage & I/O.

Slides:



Advertisements
Similar presentations
Faculty of Information Technology Department of Computer Science Computer Organization Chapter 7 External Memory Mohammad Sharaf.
Advertisements

Magnetic Disk Magnetic disks are the foundation of external memory on virtually all computer systems. A disk is a circular platter constructed of.
Computer Organization and Architecture
CSCE430/830 Computer Architecture
+ CS 325: CS Hardware and Software Organization and Architecture Internal Memory.
Agenda Types of Storage Media Semiconductor ROM and RAM Magnetic Tapes, Floppy Disks, and Hard Disks Optical CS C446 Data Storage Technologies & Networks.
Input/OutputCS510 Computer ArchitecturesLecture Lecture 16 Input and Output Devices and Systems.
CSE521: Introduction to Computer Architecture Mazin Yousif I/O Subsystem RAID (Redundant Array of Independent Disks)
CSCE 212 Chapter 8 Storage, Networks, and Other Peripherals Instructor: Jason D. Bakos.
RAID Technology. Use Arrays of Small Disks? 14” 10”5.25”3.5” Disk Array: 1 disk design Conventional: 4 disk designs Low End High End Katz and Patterson.
CS 430 – Computer Architecture Disks
CS61C L40 I/O: Disks (1) Ho, Fall 2004 © UCB TA Casey Ho inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures Lecture 39 I/O : Disks Microsoft.
Disk Storage SystemsCSCE430/830 Disk Storage Systems CSCE430/830 Computer Architecture Lecturer: Prof. Hong Jiang Courtesy of Yifeng Zhu (U. Maine) Fall,
Computer ArchitectureFall 2007 © November 28, 2007 Karem A. Sakallah Lecture 24 Disk IO and RAID CS : Computer Architecture.
CS61C L16 Disks © UC Regents 1 CS61C - Machine Structures Lecture 16 - Disks October 20, 2000 David Patterson
Other Disk Details. 2 Disk Formatting After manufacturing disk has no information –Is stack of platters coated with magnetizable metal oxide Before use,
1 Lecture 26: Storage Systems Topics: Storage Systems (Chapter 6), other innovations Final exam stats:  Highest: 95  Mean: 70, Median: 73  Toughest.
ENGS 116 Lecture 171 Introduction to I/O Vincent Berk November 19, 2008 Reading for today: Sections 4.4 – 4.9 Reading for Monday: Sections 6.1 – 6.4 Reading.
First magnetic disks, the IBM 305 RAMAC (2 units shown) introduced in One platter shown top right. A RAMAC stored 5 million characters on inch.
April 20, 2004 Prof. Andreas Savvides Spring 2004
CS61C L40 I/O: Disks (1) Garcia, Fall 2004 © UCB Lecturer PSOE Dan Garcia inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures.
Computer ArchitectureFall 2008 © November 12, 2007 Nael Abu-Ghazaleh Lecture 24 Disk IO.
Disk Technologies. Magnetic Disks Purpose: – Long-term, nonvolatile, inexpensive storage for files – Large, inexpensive, slow level in the memory hierarchy.
Cs 61C L15 Disks.1 Patterson Spring 99 ©UCB CS61C Anatomy of I/O Devices: Magnetic Disks Lecture 15 March 10, 1999 Dave Patterson (http.cs.berkeley.edu/~patterson)
S.1 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output Cache Main Memory Secondary Memory (Disk)
12/3/2004EE 42 fall 2004 lecture 391 Lecture #39: Magnetic memory storage Last lecture: –Dynamic Ram –E 2 memory This lecture: –Future memory technologies.
DAP Fall.‘00 ©UCB 1 Storage Devices and RAID Professor David A. Patterson Computer Science 252 Fall 2000.
CS 61C L41 I/O Disks (1) Garcia, Spring 2004 © UCB Lecturer PSOE Dan Garcia inst.eecs.berkeley.edu/~cs61c CS61C : Machine.
* Definition of -RAM (random access memory) :- -RAM is the place in a computer where the operating system, application programs & data in current use.
Secondary Storage Unit 013: Systems Architecture Workbook: Secondary Storage 1G.
CS 61C: Great Ideas in Computer Architecture (Machine Structures) Lecture 39: IO Disks Instructor: Dan Garcia 1.
Redundant Array of Inexpensive Disks (RAID). Redundant Arrays of Disks Files are "striped" across multiple spindles Redundancy yields high data availability.
Lecture on Electronic Memories. What Is Electronic Memory? Electronic device that stores digital information Types –Volatile v. non-volatile –Static v.
Eng. Mohammed Timraz Electronics & Communication Engineer University of Palestine Faculty of Engineering and Urban planning Software Engineering Department.
Storage & Peripherals Disks, Networks, and Other Devices.
Lecture 4 1 Reliability vs Availability Reliability: Is anything broken? Availability: Is the system still available to the user?
CS 352 : Computer Organization and Design University of Wisconsin-Eau Claire Dan Ernst Storage Systems.
RAID Storage EEL4768, Fall 2009 Dr. Jun Wang Slides Prepared based on D&P Computer Architecture textbook, etc.
CPE 631: Storage Electrical and Computer Engineering University of Alabama in Huntsville Aleksandar Milenkovic
L/O/G/O External Memory Chapter 3 (C) CS.216 Computer Architecture and Organization.
1 IKI20210 Pengantar Organisasi Komputer Kuliah No. 21: Peripheral 29 Nopember 2002 Bobby Nazief Johny Moningka
N-Tier Client/Server Architectures Chapter 4 Server - RAID Copyright 2002, Dr. Ken Hoganson All rights reserved. OS Kernel Concept RAID – Redundant Array.
I/O – Chapter 8 Introduction Disk Storage and Dependability – 8.2 Buses and other connectors – 8.4 I/O performance measures – 8.6.
1 Chapter 7: Storage Systems Introduction Magnetic disks Buses RAID: Redundant Arrays of Inexpensive Disks.
Disk Access. DISK STRUCTURE Sector: Smallest unit of data transfer from/to disk; 512B 2/4/8 adjacent sectors transferred together: Blocks Read/write heads.
Chapter 5 Internal Memory. Semiconductor Memory Types.
Lecture 13 Main Memory Computer Architecture COE 501.
Main Memory CS448.
+ CS 325: CS Hardware and Software Organization and Architecture Memory Organization.
Memory Cell Operation.
Computer Memory Storage Decoding Addressing 1. Memories We've Seen SIMM = Single Inline Memory Module DIMM = Dual IMM SODIMM = Small Outline DIMM RAM.
Disk Storage SystemsCSCE430/830 Disk Storage Systems CSCE430/830 Computer Architecture Lecturer: Prof. Hong Jiang Courtesy of Yifeng Zhu (U. Maine) Fall,
Csci 136 Computer Architecture II – IO and Storage Systems Xiuzhen Cheng
CS 61C: Great Ideas in Computer Architecture Dependability and RAID Lecture 25 Instructors: John Wawrzynek & Vladimir Stojanovic
Chapter 8 Secondary Memory. Topics Types of External Memory Magnetic Disk Optical Magnetic Tape.
1 Adapted from UC Berkeley CS252 S01 Lecture 18: Reducing Cache Hit Time and Main Memory Design Virtucal Cache, pipelined cache, cache summary, main memory.
1 Lecture 27: Disks Today’s topics:  Disk basics  RAID  Research topics.
Abstract Increases in CPU and memory will be wasted if not matched by similar performance in I/O SLED vs. RAID 5 levels of RAID and respective cost/performance.
Indira Gandhi National Open University presents. A Video Lecture Course: Computer Platforms.
CS 61C: Great Ideas in Computer Architecture Dependability - ECC Nicholas Weaver & Vladimir Stojanovic 1.
LECTURE 13 I/O. I/O CANNOT BE IGNORED Assume a program requires 100 seconds, 90 seconds for main memory, 10 seconds for I/O. Assume main memory access.
1 Components of the Virtual Memory System  Arrows indicate what happens on a lw virtual address data physical address TLB page table memory cache disk.
CMSC 611: Advanced Computer Architecture I/O & Storage Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
Module: Storage Systems
Vladimir Stojanovic & Nicholas Weaver
Lecture 13 I/O.
Lecture 28: Reliability Today’s topics: GPU wrap-up Disk basics RAID
BIC 10503: COMPUTER ARCHITECTURE
Presentation transcript:

Graduate Computer Architecture I Lecture 13: Storage & I/O

2 - CSE/ESE 560M – Graduate Computer Architecture I Row and Column Address together: –Select 1 bit a time rowdecoderrowdecoder row address Column Selector & I/O Circuits Column Address data RAM Cell Array word (row) select bit (data) lines Each intersection represents a 1-T DRAM Cell Classical DRAM Organization

3 - CSE/ESE 560M – Graduate Computer Architecture I 1-T Memory Cell (DRAM) Write: –1. Drive bit line –2.. Select row Read: –1. Precharge bit line to Vdd/2 –2.. Select row –3. Cell and bit line share charges Very small voltage changes on the bit line –4. Sense (fancy sense amp) Can detect changes of ~1 million electrons –5. Write: restore the value Refresh –1. Just do a dummy read to every cell. row select bit

4 - CSE/ESE 560M – Graduate Computer Architecture I Main Memory Performance Time Access Time Cycle Time DRAM (Read/Write) Cycle Time > DRAM (Read/Write) Access Time DRAM (Read/Write) Cycle Time : –How frequent can you initiate an access? –Analogy: A little kid can only ask his father for money on Saturday DRAM (Read/Write) Access Time: –How quickly will you get what you want once you initiate an access? –Analogy: As soon as he asks, his father will give him the money DRAM Bandwidth Limitation analogy: –What happens if he runs out of money on Wednesday?

5 - CSE/ESE 560M – Graduate Computer Architecture I Access Pattern without Interleaving: Start Access for D1 CPUMemory Start Access for D2 D1 available Access Pattern with 4-way Interleaving: Access Bank 0 Access Bank 1 Access Bank 2 Access Bank 3 We can Access Bank 0 again CPU Memory Bank 1 Memory Bank 0 Memory Bank 3 Memory Bank 2 Increasing Bandwidth - Interleaving

6 - CSE/ESE 560M – Graduate Computer Architecture I Main Memory Performance Simple –CPU, Cache, Bus, Memory same width (32 bits) Interleaved –CPU, Cache, Bus 1 word: Memory N Modules (4 Modules); example is word interleaved Wide –CPU/Mux 1 word; Mux/Cache, Bus, Memory N words (Alpha: 64 bits & 256 bits)

7 - CSE/ESE 560M – Graduate Computer Architecture I Main Memory Performance address Bank address Bank address Bank address Bank Timing model –1 to send address, –4 for access time, 10 cycle time, 1 to send data –Cache Block is 4 words Simple M.P. = 4 x (1+10+1) = 48 Wide M.P. = = 12 Interleaved M.P. = =15

8 - CSE/ESE 560M – Graduate Computer Architecture I DRAM History DRAMs: capacity +60%/yr, cost –30%/yr –2.5X cells/area, 1.5X die size in ­3 years ‘98 DRAM fab line costs $2B –DRAM only: density, leakage v. speed Rely on increasing no. of computers & memory per computer (60% market) –SIMM or DIMM is replaceable unit –computers use any generation DRAM Commodity, second source industry –high volume, low profit, conservative –Little organization innovation in 20 years Order of importance –(1) Cost/bit (2) Capacity (3) Performance –RAMBUS: 10X BW, +30% cost  little impact

9 - CSE/ESE 560M – Graduate Computer Architecture I DRAM Crossroads After 20 years of 4X every 3 years, running into wall? (64Mb - 1 Gb) How can keep $1B fab lines full if buy fewer DRAMs per computer? Cost/bit –30%/yr if stop 4X/3 yr? What will happen to $40B/yr DRAM industry?

10 - CSE/ESE 560M – Graduate Computer Architecture I Tunneling Magnetic Junction Tunneling Magnetic Junction RAM (TMJ-RAM) –6x faster than DRAM –Density of DRAM, non-volatile (no refresh) –“Spintronics”: combination quantum spin and electronics –Same technology used in high-density disk-drives

11 - CSE/ESE 560M – Graduate Computer Architecture I Carbon Nanotube Memory 10,000 to 30,000 times more dense than today's Dynamic Random Access Memory (DRAM) memory chips, said Charles M. Lieber, professor of chemistry at Harvard and leader of the team that developed the nanotube technique. "It's much higher density than… will ever be attainable with silicon [wafer] technology," he said. In principle these memory arrays could also be 1,000 to 10,000 times faster than today's memory chips, he added.

12 - CSE/ESE 560M – Graduate Computer Architecture I Micro-Electro-Mechanical Systems (MEMS) Magnetic “sled” floats on array of read/write heads –Approx 250 Gbit/in 2 –Data rates: IBM: 250 MB/s w1000 heads CMU: 3.1 MB/s w 400 heads Electrostatic actuators move media around to align it with heads –Sweep sled ±50  m in < 0.5  s Capacity estimated to be in the 1-10GB in 1cm 2

13 - CSE/ESE 560M – Graduate Computer Architecture I Storage Errors Motivation: –DRAM is dense  Signals are easily disturbed –High Capacity  higher probability of failure Approach: Redundancy –Add extra information so that we can recover from errors –Can we do better than just create complete copies?

14 - CSE/ESE 560M – Graduate Computer Architecture I Error Correction Codes (ECC) Memory systems generate errors (accidentally flipped-bits) –DRAMs store very little charge per bit –“Soft” errors occur occasionally when cells are struck by alpha particles or other environmental upsets. –Less frequently, “hard” errors can occur when chips permanently fail. –Problem gets worse as memories get denser and larger Where is “perfect” memory required? –servers, spacecraft/military computers, ebay, … Memories are protected against failures with ECCs Extra bits are added to each data-word –used to detect and/or correct faults in the memory system –in general, each possible data word value is mapped to a unique “code word”. A fault changes a valid code word to an invalid one - which can be detected.

15 - CSE/ESE 560M – Graduate Computer Architecture I Simple Error Detection Coding Each data value, before it is written to memory is “tagged” with an extra bit to force the stored word to have even parity: Each word, as it is read from memory is “checked” by finding its parity (including the parity bit). Parity Bit b7b6b5b4b3b2b1b0pb7b6b5b4b3b2b1b0p + b7b6b5b4b3b2b1b0pb7b6b5b4b3b2b1b0p + c A non-zero parity indicates an error occurred: –two errors (on different bits) is not detected (nor any even number of errors) –odd numbers of errors are detected. What is the probability of multiple simultaneous errors? Extend the idea for Hamming ECC

16 - CSE/ESE 560M – Graduate Computer Architecture I Input and Output CPU Performance: 60% per year I/O system performance limited by mechanical delays (disk I/O) –< 10% per year (IO per sec or MB per sec) Amdahl's Law: system speed-up limited by the slowest part! –10% IO & 10x CPU => 5x Performance (lose 50%) –10% IO & 100x CPU => 10x Performance (lose 90%) I/O bottleneck: –Diminishing fraction of time in CPU –Diminishing value of faster CPUs

17 - CSE/ESE 560M – Graduate Computer Architecture I Processor Cache Memory - I/O Bus Main Memory I/O Controller Disk I/O Controller I/O Controller Graphics Network interrupts I/O Systems

18 - CSE/ESE 560M – Graduate Computer Architecture I Disk Capacity now doubles every 18 months; before 1990 every 36 months Today: Processing Power Doubles Every 18 months Today: Memory Size Doubles Every 18 months(4X/3yr) Today: Disk Capacity Doubles Every 18 months Disk Positioning Rate (Seek + Rotate) Doubles Every Ten Years! The I/O GAP The I/O GAP Technology Trends

19 - CSE/ESE 560M – Graduate Computer Architecture I 1956 IBM Ramac — early 1970s Winchester –Developed for mainframe computers, proprietary interfaces –Steady shrink in form factor: 27 in. to 14 in. 1970s developments –5.25 inch floppy disk formfactor (microcode into mainframe) –early emergence of industry standard disk interfaces ST506, SASI, SMD, ESDI Early 1980s –PCs and first generation workstations Mid 1980s –Client/server computing –Centralized storage on file server accelerates disk downsizing: 8 inch to 5.25 inch –Mass market disk drives become a reality industry standards: SCSI, IPI, IDE 5.25 inch drives for standalone PCs, End of proprietary interfaces Historical Perspective

20 - CSE/ESE 560M – Graduate Computer Architecture I Disk History 1973: 1. 7 Mbit/sq. in 140 MBytes 1979: 7. 7 Mbit/sq. in 2,300 MBytes

21 - CSE/ESE 560M – Graduate Computer Architecture I Historical Perspective Late 1980s/Early 1990s: –Laptops, notebooks, (palmtops) –3.5 inch, 2.5 inch, (1.8 inch form-factors) –Form-factor plus capacity drives market, not so much performance Recently Bandwidth improving at 40%/ year –Challenged by DRAM, flash RAM in PCMCIA cards Still expensive, Intel promises but doesn’t deliver Unattractive MBytes per cubic inch –Optical disk fails on performance Finds niche (CD ROM)

22 - CSE/ESE 560M – Graduate Computer Architecture I Disk History 1989: 63 Mbit/sq. in 60,000 MBytes 1997: 1450 Mbit/sq. in 2300 MBytes 1997: 3090 Mbit/sq. in 8100 MBytes

23 - CSE/ESE 560M – Graduate Computer Architecture I Disk Performance Model /Trends Capacity + 100%/year (2X / 1.0 yrs) Transfer rate (BW) + 40%/year (2X / 2.0 yrs) Rotation + Seek time – 8%/ year (1/2 in 10 yrs) MB/$ > 100%/year (2X / <1.5 yrs) Fewer chips + areal density

24 - CSE/ESE 560M – Graduate Computer Architecture I Actuator Arm Head Platters (12) { Spindle Photo of Disk Head, Arm, Actuator

25 - CSE/ESE 560M – Graduate Computer Architecture I Coil for writing Special sensitivity of Disk head comes from “Giant Magneto-Resistive effect” or (GMR) IBM is (was) leader in this technology –Same technology as TMJ-RAM breakthrough Nano-layered Disk Heads

26 - CSE/ESE 560M – Graduate Computer Architecture I Disk Device Terminology Platter Outer Track Inner Track Sector Actuator HeadArm Several platters –information recorded magnetically on both surfaces (usually) Bits recorded –in tracks, which in turn divided into sectors (e.g., 512 Bytes) Actuator –moves head (end of arm,1/surface) over track (“seek”), select surface, wait for sector rotate under head, then read or write “Cylinder” - all tracks under heads

27 - CSE/ESE 560M – Graduate Computer Architecture I CapBPITPIBPI*TPIData Xfer Access Technology(MB)(Million)(KByte/s) Time Conventional Tape: Cartridge (.25") minutes IBM 3490 (.5") seconds Helical Scan Tape: Video (8mm) secs DAT (4mm) secs Magnetic & Optical Disk: Hard Disk (5.25") ms IBM 3390 (10.5") ms Sony MO (5.25") ms Alternative Data Storage: Early 1990s

28 - CSE/ESE 560M – Graduate Computer Architecture I Tape wear out: –Helical 100s of passes to 1000s for longitudinal Head wear out: –2000 hours for helical Both must be accounted for in economic / reliability model Long rewind, eject, load, spin-up times; not inherent, just no need in marketplace (so far) Designed for archival Current Drawbacks to Tape

29 - CSE/ESE 560M – Graduate Computer Architecture I Memory Cost 2005 Memory Modules SRAM2 MB$26$13000/GB DRAM1 GB$90$90/GB Flash1 GB$50$50/GB Magnetic Disks 3.5”200 GB$150$0.75/GB 2.5”70 GB$100$1.43/GB Optical Disks 5.25”4.6 GB$2$0.43/GB

30 - CSE/ESE 560M – Graduate Computer Architecture I 14” 10”5.25”3.5” Disk Array: 1 disk design Conventional: 4 disk designs Low End High End Disk Product Families Disk Arrays

31 - CSE/ESE 560M – Graduate Computer Architecture I Data Capacity Volume Power Data Rate I/O Rate MTTF Cost IBM 3390 (K) 20 GBytes 97 cu. ft. 3 KW 15 MB/s 600 I/Os/s 250 KHrs $250K IBM 3.5" MBytes 0.1 cu. ft. 11 W 1.5 MB/s 55 I/Os/s 50 KHrs $2K x70 23 GBytes 11 cu. ft. 1 KW 120 MB/s 3900 IOs/s ??? Hrs $150K Disk Arrays have potential for large data and I/O rates high MB per cu. ft., high MB per KW reliability? Large Disk  Array of Small Disks

32 - CSE/ESE 560M – Graduate Computer Architecture I Hot spares support reconstruction in parallel with access: very high media availability can be achieved Hot spares support reconstruction in parallel with access: very high media availability can be achieved Array Reliability Reliability of N disks = Reliability of 1 Disk ÷ N –50,000 Hours ÷ 70 disks = 700 hours Disk System Life: Drops from 6 years to 1 month! Arrays (without redundancy) too unreliable to be useful!

33 - CSE/ESE 560M – Graduate Computer Architecture I Files are "striped" across multiple spindles Redundancy yields high data availability Disks will fail Contents reconstructed from data redundantly stored in the array Capacity penalty to store it Bandwidth penalty to update Mirroring/Shadowing (high capacity cost) Horizontal Hamming Codes (overkill) Parity & Reed-Solomon Codes Failure Prediction (no capacity overhead!) VaxSimPlus — Technique is controversial Techniques: Redundant Arrays of Disks

34 - CSE/ESE 560M – Graduate Computer Architecture I P logical record Striped physical records Parity computed across recovery group to protect against hard disk failures 33% capacity cost for parity in this configuration wider arrays reduce capacity costs, decrease expected availability, increase reconstruction time Arms logically synchronized, spindles rotationally synchronized logically a single high capacity, high transfer rate disk Targeted for high bandwidth applications: Scientific, Image Processing RAID 3: Parity Disk

35 - CSE/ESE 560M – Graduate Computer Architecture I A logical write becomes four physical I/Os Independent writes possible because of interleaved parity Reed-Solomon Codes ("Q") for protection during reconstruction A logical write becomes four physical I/Os Independent writes possible because of interleaved parity Reed-Solomon Codes ("Q") for protection during reconstruction D0D1D2 D3 P D4D5D6 P D7 D8D9P D10 D11 D12PD13 D14 D15 PD16D17 D18 D19 D20D21D22 D23 P Disk Columns Increasing Logical Disk Addresses Stripe Unit Targeted for mixed applications RAID 5+: High I/O Rate Parity

36 - CSE/ESE 560M – Graduate Computer Architecture I IRAM Combine Processor with Memory in One Chip –Large Bandwidth –Bandwidth of nearly 1000 gigabits per second (32K bits in 50 ns) –A hundredfold increase over the fastest computers today. –The fastest programs will keep most memory accesses within a single IRAM, rewarding compact representations of code and data