Lecture 11: I/O Management and Disk Scheduling. Categories of I/O Devices Human readable –Used to communicate with the user –Printers –Video display terminals.

Slides:



Advertisements
Similar presentations
I/O Management and Disk Scheduling
Advertisements

I/O Management and Disk Scheduling
1 Chapter 11 I/O Management and Disk Scheduling Patricia Roy Manatee Community College, Venice, FL ©2008, Prentice Hall Operating Systems: Internals and.
I/O Management and Disk Scheduling Chapter 11. I/O Driver OS module which controls an I/O device hides the device specifics from the above layers in the.
Chapter 4 Device Management and Disk Scheduling DEVICE MANAGEMENT Content I/O device overview I/O device overview I/O organization and architecture I/O.
I/O Management and Disk Scheduling
Chapter 11 I/O Management and Disk Scheduling
I/O Management and Disk Scheduling
Chapter 11 I/O Management and Disk Scheduling
Categories of I/O Devices
CS 6560: Operating Systems Design
Chapter 11 I/O Management and Disk Scheduling
Day 30 I/O Management and Disk Scheduling. I/O devices Vary in many ways Vary in many ways –Data rate –Application –Complexity of control –Unit of transfer.
Input/Output Management and Disk Scheduling
Lecture 17 I/O Optimization. Disk Organization Tracks: concentric rings around disk surface Sectors: arc of track, minimum unit of transfer Cylinder:
Based on the slides supporting the text
1 Chapter 11 I/O Management and Disk Scheduling –Operating System Design Issues –I/O Buffering –Disk Scheduling –Disk Cache.
Computer System Overview
Chapter 1 and 2 Computer System and Operating System Overview
I/O Management and Disk Scheduling (Chapter 10) Perhaps the messiest aspect of operating system design is input/output A wide variety of devices and many.
Device Management.
Computer Organization and Architecture
I/O Management and Disk Scheduling Chapter 11. Categories: For Human interaction : Printers, terminals, keyboard, mouse Machine readable: Disks, Sensors,
Disks CS 416: Operating Systems Design, Spring 2001 Department of Computer Science Rutgers University
1 I/O Management in Representative Operating Systems.
1 Today I/O Systems Storage. 2 I/O Devices Many different kinds of I/O devices Software that controls them: device drivers.
RAID Ref: Stallings. Introduction The rate in improvement in secondary storage performance has been considerably less than the rate for processors and.
1 Recitation 8 Disk & File System. 2 Disk Scheduling Disks are at least four orders of magnitude slower than main memory –The performance of disk I/O.
I/O Management and Disk Scheduling
Chapter 11 I/O Management and Disk Scheduling Patricia Roy Manatee Community College, Venice, FL ©2008, Prentice Hall Operating Systems: Internals and.
Computer Systems Overview. Page 2 W. Stallings: Operating Systems: Internals and Design, ©2001 Operating System Exploits the hardware resources of one.
1 Computer System Overview Chapter 1. 2 n An Operating System makes the computing power available to users by controlling the hardware n Let us review.
Computer System Overview Chapter 1. Operating System Exploits the hardware resources of one or more processors Provides a set of services to system users.
Chapter 1 Computer System Overview Dave Bremer Otago Polytechnic, N.Z. ©2008, Prentice Hall Operating Systems: Internals and Design Principles, 6/E William.
Disk Structure Disk drives are addressed as large one- dimensional arrays of logical blocks, where the logical block is the smallest unit of transfer.
1 I/O Management and Disk Scheduling Chapter Categories of I/O Devices Human readable Used to communicate with the user Printers Video display terminals.
1 IO Management and Disk Scheduling Chapter Categories of I/O Devices n Human readable u used to communicate with the user u video display terminals.
1 I/O Management and Disk Scheduling Chapter
1 I/O Management and Disk Scheduling Chapter Categories of I/O Devices Human readable  Used to communicate with the user  Printers  Video display.
© Janice Regan, CMPT 300, May CMPT 300 Introduction to Operating Systems Principles of I/0 hardware.
1 I/O Management and Disk Scheduling Chapter Categories of I/O Devices Human readable Used to communicate with the user Printers Video display terminals.
CE Operating Systems Lecture 20 Disk I/O. Overview of lecture In this lecture we will look at: Disk Structure Disk Scheduling Disk Management Swap-Space.
I/O Management and Disk Structure Introduction to Operating Systems: Module 14.
1 I/O Management and Disk Scheduling Chapter Categories of I/O Devices Human readable –Used to communicate with the user –Printers –Video display.
I/O and Disk Scheduling CS Spring Overview Review of I/O Techniques I/O Buffering Disk Geometry Disk Scheduling Algorithms RAID.
O VERVIEW OF M ASS S TORAGE S TRUCTURE Magnetic disks provide bulk of secondary storage of modern computers Drives rotate at 60 to 250 times per second.
Fall 2000M.B. Ibáñez Lecture 25 I/O Systems. Fall 2000M.B. Ibáñez Categories of I/O Devices Human readable –used to communicate with the user –video display.
I/O Management and Disk Scheduling Chapter 11. Disk Performance Parameters To read or write, the disk head must be positioned at the desired track and.
OPERATING SYSTEMS IO SYSTEMS. Categories of I/O Devices Human readable –Used to communicate with the user –Printers –Video display terminals Display Keyboard.
I/O Management and Disk Scheduling Chapter 11. Categories of I/O Devices Human readable Used to communicate with the user  Printers, Display terminals,
1 I/O Management and Disk Scheduling Chapter I/O Devices Can be group into three categories: 1.Human readable Used to communicate with the computer.
Chapter 14: Mass-Storage Systems Disk Structure. Disk Scheduling. RAID.
Chapter 101 I/O Management and Disk Scheduling Chapter 11.
Part IV I/O System Chapter 12: Mass Storage Structure.
Chapter 11 I/O Management and Disk Scheduling Dave Bremer Otago Polytechnic, NZ ©2008, Prentice Hall Operating Systems: Internals and Design Principles,
INPUT OUTPUT MANAGEMENT Prepared By: Dr. Vipul Vekariya.
Computer Systems Overview. Lecture 1/Page 2AE4B33OSS W. Stallings: Operating Systems: Internals and Design, ©2001 Operating System Exploits the hardware.
1 Computer System Overview Chapter 1. 2 Operating System Exploits the hardware resources of one or more processors Provides a set of services to system.
1 Chapter 11 I/O Management and Disk Scheduling Patricia Roy Manatee Community College, Venice, FL ©2008, Prentice Hall Operating Systems: Internals and.
I/O Management and Disk Scheduling
Multiple Platters.
RAID, Programmed I/O, Interrupt Driven I/O, DMA, Operating System
I/O System Chapter 5 Designed by .VAS.
Operating Systems Disk Scheduling A. Frank - P. Weisberg.
Operating System I/O System Monday, August 11, 2008.
Overview Continuation from Monday (File system implementation)
Chapter 11 I/O Management and Disk Scheduling
Disk Scheduling The operating system is responsible for using hardware efficiently — for the disk drives, this means having a fast access time and disk.
Operating Systems Disk Scheduling A. Frank - P. Weisberg.
Presentation transcript:

Lecture 11: I/O Management and Disk Scheduling

Categories of I/O Devices Human readable –Used to communicate with the user –Printers –Video display terminals Display Keyboard Mouse

Categories of I/O Devices Machine readable –Used to communicate with electronic equipment –Disk and tap drives –Sensors –Controllers –Actuators

Categories of I/O Devices Communication –Used to communicate with remote devices –Digital line drivers –Modems

Differences in I/O Devices Data rate –May be differences of several orders of magnitude between the data transfer rates

Differences in I/O Devices Application –Disk used to store files requires file- management software –Disk used to store virtual memory pages needs special hardware and software to support it –Terminal used by system administrator may have a higher priority

Differences in I/O Devices Complexity of control Unit of transfer –Data may be transferred as a stream of bytes for a terminal or in larger blocks for a disk Data representation –Encoding schemes Error conditions –Devices respond to errors differently

Differences in I/O Devices Programmed I/O –Process is busy-waiting for the operation to complete Interrupt-driven I/O –I/O command is issued –Processor continues executing instructions –I/O module sends an interrupt when done

Techniques for Performing I/O Direct Memory Access (DMA) –DMA module controls exchange of data between main memory and the I/O device –Processor interrupted only after entire block has been transferred

Evolution of the I/O Function Processor directly controls a peripheral device Controller or I/O module is added –Processor uses programmed I/O without interrupts –Processor does not need to handle details of external devices

Evolution of the I/O Function Controller or I/O module with interrupts –Processor does not spend time waiting for an I/O operation to be performed Direct Memory Access –Blocks of data are moved into memory without involving the processor –Processor involved at beginning and end only

Evolution of the I/O Function I/O module is a separate processor I/O processor –I/O module has its own local memory –Its a computer in its own right

Direct Memory Access Takes control of the system form the CPU to transfer data to and from memory over the system bus Cycle stealing is used to transfer data on the system bus The instruction cycle is suspended so data can be transferred The CPU pauses one bus cycle No interrupts occur –No need to save context

DMA

Cycle stealing causes the CPU to execute more slowly Number of required busy cycles can be cut by integrating the DMA and I/O functions Path between DMA module and I/O module that does not include the system bus

DMA

Operating System Design Issues Efficiency –Most I/O devices extremely slow compared to main memory –Use of multiprogramming allows for some processes to be waiting on I/O while another process executes –I/O cannot keep up with processor speed –Swapping is used to bring in additional ready processes, which is an I/O operation

Operating System Design Issues Generality –Desirable to handle all I/O devices in a uniform manner –Hide most of the details of device I/O in lower-level routines so that processes and upper levels see devices in general terms such as read, write, open, close, lock, unlock

I/O Buffering Reasons for buffering –Processes must wait for I/O to complete before proceeding –Certain pages must remain in main memory during I/O

I/O Buffering Block-oriented –Information is stored in fixed sized blocks –Transfers are made a block at a time –Used for disks and tapes Stream-oriented –Transfer information as a stream of bytes –Used for terminals, printers, communication ports, mouse, and most other devices that are not secondary storage

Single Buffer Operating system assigns a buffer in main memory for an I/O request Block-oriented –Input transfers made to buffer –Block moved to user space when needed –Another block is moved into the buffer Read ahead

I/O Buffering

Single Buffer Block-oriented –User process can process one block of data while next block is read in –Swapping can occur since input is taking place in system memory, not user memory –Operating system keeps track of assignment of system buffers to user processes

Single Buffer Stream-oriented –Used a line at time –User input from a terminal is one line at a time with carriage return signaling the end of the line –Output to the terminal is one line at a time

Double Buffer Use two system buffers instead of one A process can transfer data to or from one buffer while the operating system empties or fills the other buffer

Circular Buffer More than two buffers are used Each individual buffer is one unit in a circular buffer Used when I/O operation must keep up with process

I/O Buffering

I/O Driver OS module that controls an I/O device hides the device specifics from the above layers in the OS/kernel translates logical I/O into device I/O (logical disk blocks into {track, head, sector}) performs data buffering and scheduling of I/O operations structure: several synchronous entry points (device initialization, queue I/O requests, state control, read/write) and an asynchronous entry point (to handle interrupts)

Typical Driver Structure driver_strategy(request) { if (empty(request-queue)) driver_start(request) else add(request, request-queue) block_current_process; reschedule() } driver_start(request) { current_request= request; start_dma(request); } driver_ioctl(request) { } driver_init() { } driver_interrupt(state) /* asynchronous part */ { if (state==ERROR) && (retries++<MAX) { driver_start(current_request); return; } add_current_process_to_active_queue if (! (empty(request_queue)) driver_start(get_next(request_queue)) }

User to Driver Control Flow user kernel read, write, ioctl special file ordinary file File System Buffer Cache block device character device Character queue driver_read/writedriver-strategy

I/O Buffering before an I/O request is placed, the source/destination of the I/O transfer must be locked in memory I/O buffering: data is copied from user space to kernel buffers, which are pinned to memory buffer cache: a buffer in main memory for disk sectors character queue: follows the producer/consumer model (characters in the queue are read once) unbuffered I/O to/from disk (block device): VM paging

Buffer Cache when an I/O request is made for a sector, the buffer cache is checked first if it is missing from the cache, it is read into the buffer cache from the disk exploits locality of reference as any other cache usually, replacements are done in chunks (a whole track can be written back at once to minimize seek time) replacement policies are global and controlled by the kernel

Replacement Policies buffer cache organized like a stack: replace from the bottom LRU: replace the block that has been in the cache longest with no reference to it (on reference a block is moved to the top of the stack) LFU: replace the block with the fewest references (counters that are incremented on reference; blocks move accordingly) frequency-based replacement: define a new section on the top of the stack, counter is unchanged while the block is in the new section

Least Recently Used The block that has been in the cache the longest with no reference to it is replaced The cache consists of a stack of blocks Most recently referenced block is on the top of the stack When a block is referenced or brought into the cache, it is placed on the top of the stack

Least Frequently Used The block that has experienced the fewest references is replaced A counter is associated with each block Counter is incremented each time the block is accessed Block with smallest count is selected for replacement Some blocks may be referenced many times in a short period of time and then not needed any more

Application-Controlled File Caching two-level block replacement: responsibility is split between kernel and user level the global allocation policy performed by the kernel, which decides which process will free a block the block replacement policy decided by the user: kernel provides the candidate block as a hint to the process the process can overrule the kernel’s choice by suggesting an alternative block the suggested block is replaced by the kernel examples of alternative replacement policy: most- recently used (MRU)

Sound Kernel-User Cooperation oblivious processes should do no worse than under LRU foolish processes should not hurt other processes smart processes should perform better than LRU whenever possible and they should never perform worse if kernel selects block A and user chooses B instead, the kernel swaps the position of A and B in the LRU list and places B in a “placeholder” that points to A (kernel’s choice) if the user process misses on B (i.e. he made a bad choice), and B is found in the placeholder, then the block pointed to by the placeholder is chosen (prevents hurting other processes)

Disk Performance Parameters To read or write, the disk head must be positioned at the desired track and at the beginning of the desired sector Seek time –time it takes to position the head at the desired track Rotational delay or rotational latency –time its takes for the beginning of the sector to reach the head

Disk Performance Parameters Access time –Sum of seek time and rotational delay –The time it takes to get in position to read or write Data transfer occurs as the sector moves under the head

Disk I/O Performance disks are at least four orders of magnitude slower than the main memory the performance of disk I/O is vital for the performance of the computer system as a whole disk performance parameters seek time (to position the head at the track): 20 ms rotational delay (to reach the sector): 8.3 ms transfer time: 1-2 MB/sec access time (seek time+ rotational delay) >> transfer time for a sector the order in which sectors are read matters a lot

Disk Scheduling Policies Seek time is the reason for differences in performance For a single disk, there will be a number of I/O requests If requests are selected randomly, we will get the worst possible performance

Disk Scheduling Policies First-in, first-out (FIFO) –Process request sequentially –Fair to all processes –Approaches random scheduling in performance if there are many processes

Disk Scheduling Policies Priority –Goal is not to optimize disk use but to meet other objectives –Short batch jobs may have higher priority –Provide good interactive response time

Disk Scheduling Policies Last-in, first-out –Good for transaction processing systems The device is given to the most recent user in order to have little arm movement –Possibility of starvation since a job may never regain the head of the line

Disk Scheduling Policies Shortest Service Time First –Select the disk I/O request that requires the least movement of the disk arm from its current position –Always choose the minimum seek time

Disk Scheduling Policies SCAN –Arm moves in one direction only, satisfying all outstanding requests until it reaches the last track in that direction –Direction is reversed

Disk Scheduling Policies C-SCAN –Restricts scanning to one direction only –When the last track has been visited in one direction, the arm is returned to the opposite end of the disk and the scan begins again

Disk Scheduling Policies N-step-SCAN –Segments the disk request queue into subqueues of length N –Subqueues are process one at a time, using SCAN –New requests added to other queue when queue is processed FSCAN –Two queues –One queue is empty for new request

Disk Scheduling Policies usually based on the position of the requested sector rather than according to the process priority shortest-service-time-first (SSTF): pick the request that requires the least movement of the head SCAN (back and forth over disk): good distribution C-SCAN(one way with fast return):lower service variability but head may not be moved for a considerable period of time N-step SCAN: scan of N records at a time by breaking the request queue in segments of size at most N FSCAN: uses two subqueues, during a scan one queue is consumed while the other one is produced

RAID Redundant Array of Independent Disks (RAID) idea: replace large-capacity disks with multiple smaller-capacity drives to improve the I/O performance RAID is a set of physical disk drives viewed by the OS as a single logical drive data are distributed across physical drives in a way that enables simultaneous access to data from multiple drives redundant disk capacity is used to compensate the increase in the probability of failure due to multiple drives size RAID levels (design architectures)

RAID Level 0 does not include redundancy data is stripped across the available disks disk is divided into strips strips are mapped round-robin to consecutive disks a set of consecutive strips that map exactly one strip to each array member is called stripe strip 0 strip 3 strip 2 strip 1 strip 7strip 6strip 5strip 4...

RAID Level 1 redundancy achieved by duplicating all the data each logical disk is mapped to two separate physical disks so that every disk has a mirror disk that contains the same data a read can be serviced by either of the two disks that contains the requested data (improved performance over RAID 0 if reads dominate) a write request must be done on both disks but can be done in parallel recovery is simple but cost is high strip 0 strip 1 strip 0 strip 1 strip 3strip 2strip 3strip 2...

RAID Levels 2 and 3... parallel access: all disks participate in every I/O request small strips (byte or word size) RAID 2: error correcting code (Hamming) is calculated across corresponding bits on each data disk and stored on log(data) parity disks; necessary only if error rate is high RAID 3: a single redundant disk which keeps the parity bit P(i) = X2(i) + X1(i) + X0(i) in the event of failure, data can be reconstructed but only one request at the time can be satisfied b0 b1 b2P(b) X2(i) = P(i) + X1(i) + X0(i)

RAID Levels 4 and 5 strip 0 P(0-2) strip 2 strip 1 P(3-5)strip 5strip 4strip 3 independent access: each disk operates independently, multiple I/O request can be satisfied in parallel large strips RAID 4: for small writes: 2 reads + 2 writes example: if write performed only on strip 0: P’(i) = X2(i) + X1(i) + X0’1(i) = X2(i) + X1(i) + X0’(i) + X0(i) + X0(i) = P(i) + X0’(i) + X0(i) RAID 5: parity strips are distributed across all disks

UNIX SVR4 I/O