Presentation is loading. Please wait.

Presentation is loading. Please wait.

Concurrency: Mutual Exclusion and Synchronization

Similar presentations


Presentation on theme: "Concurrency: Mutual Exclusion and Synchronization"— Presentation transcript:

1 Concurrency: Mutual Exclusion and Synchronization
Unit - III Concurrency: Mutual Exclusion and Synchronization

2 Distributed Processing
Multiple Processes Operating System design is concerned with the management of processes and threads: Multiprogramming Multiprocessing Distributed Processing The central themes of operating system design are all concerned with the management of processes and threads: • Multiprogramming: The management of multiple processes within a uniprocessor system • Multiprocessing : The management of multiple processes within a multiprocessor • Distributed processing: The management of multiple processes executing on multiple, distributed computer systems. The recent proliferation of clusters is a prime example of this type of system.

3 Concurrency Arises in Three Different Contexts:
Multiple Applications invented to allow processing time to be shared among active applications Structured Applications extension of modular design and structured programming Operating System Structure OS themselves implemented as a set of processes or threads Fundamental to all of these areas, and fundamental to OS design, is concurrency. Concurrency encompasses a host of design issues, including communication among processes, sharing of and competing for resources (such as memory, files, and I/O access), synchronization of the activities of multiple processes, and allocation of processor time to processes. We shall see that these issues arise not just in multiprocessing and distributed processing environments but even in single-processor multiprogramming systems. Concurrency arises in three different contexts: • Multiple applications: Multiprogramming was invented to allow processing time to be dynamically shared among a number of active applications. • Structured applications: As an extension of the principles of modular design and structured programming, some applications can be effectively programmed as a set of concurrent processes. Operating system structure: The same structuring advantages apply to systems programs, and we have seen that operating systems are themselves often implemented as a set of processes or threads.

4 Concurrency & Shared Data
Concurrent processes may share data to support communication, info exchange,... Threads in the same process can share global address space Concurrent sharing may cause problems For example: lost updates

5 Concurrency Key Terms Key terms related to concurrency.
Table Some Key Terms Related to Concurrency

6 Principles of Concurrency
Interleaving and overlapping can be viewed as examples of concurrent processing both present the same problems In multiprogramming, the relative speed of execution of processes cannot be predicted depends on activities of other processes the way the OS handles interrupts scheduling policies of the OS At first glance, it may seem that interleaving and overlapping represent fundamentally different modes of execution and present different problems. In fact, both techniques can be viewed as examples of concurrent processing, and both present the same problems. In the case of a uniprocessor, the problems stem from a basic characteristic of multiprogramming systems: The relative speed of execution of processes cannot be predicted. It depends on the activities of other processes, the way in which the OS handles interrupts, and the scheduling policies of the OS.

7 Difficulties of Concurrency
Sharing of global resources Difficult for the OS to manage the allocation of resources optimally Difficult to locate programming errors as results are not deterministic and reproducible The following difficulties arise: 1. The sharing of global resources is fraught with peril. For example, if two processes both make use of the same global variable and both perform reads and writes on that variable, then the order in which the various reads and writes are executed is critical. An example of this problem is shown in the following subsection. 2. It is difficult for the OS to manage the allocation of resources optimally. For example, process A may request use of, and be granted control of, a particular I/O channel and then be suspended before using that channel. It may be undesirable for the OS simply to lock the channel and prevent its use by other processes; indeed this may lead to a deadlock condition, as described in Chapter 6 . 3. It becomes very difficult to locate a programming error because results are typically not deterministic and reproducible (e.g., see [LEBL87, CARR89, SHEN02] for a discussion of this point). All of the foregoing difficulties present themselves in a multiprocessor system as well, because here too the relative speed of execution of processes is unpredictable. A multiprocessor system must also deal with problems arising from the simultaneous execution of multiple processes. Fundamentally, however, the problems are the same as those for uniprocessor systems. This should become clear as the discussion proceeds.

8 The final result depends on the order of execution
Race Condition Occurs when multiple processes or threads read and write shared data items The final result depends on the order of execution the “loser” of the race is the process that updates last and will determine the final value of the variable A race condition occurs when multiple processes or threads read and write data items so that the final result depends on the order of execution of instructions in the multiple processes. Let us consider two simple examples. As a first example, suppose that two processes, P1 and P2, share the global variable a . At some point in its execution, P1 updates a to the value 1, and at some point in its execution, P2 updates a to the value 2. Thus, the two tasks are in a race to write variable a . In this example, the “loser” of the race (the process that updates last) determines the final value of a . For our second example, consider two process, P3 and P4, that share global variables b and c , with initial values b = 1 and c = 2 . At some point in its execution, P3 executes the assignment b = b + c , and at some point in its execution, P4 executes the assignment c = b + c . Note that the two processes update different variables. However, the final values of the two variables depend on the order in which the two processes execute these two assignments. If P3 executes its assignment statement first, then the final values are b = 3 and c = 5 . If P4 executes its assignment statement first, then the final values are b = 4 and c = 3 . Appendix A includes a discussion of race conditions using semaphores as an example.

9 Operating System Concerns
Design and management issues raised by the existence of concurrency: The OS must: be able to keep track of various processes allocate and de-allocate resources for each active process protect the data and physical resources of each process against interference by other processes ensure that the processes and outputs are independent of the processing speed What design and management issues are raised by the existence of concurrency? We can list the following concerns: 1. The OS must be able to keep track of the various processes. This is done with the use of process control blocks and was described in Chapter 4 . 2. The OS must allocate and de-allocate various resources for each active process. At times, multiple processes want access to the same resource. These resources include • Processor time: This is the scheduling function, discussed in Part Four. • Memory: Most operating systems use a virtual memory scheme. The topic is addressed in Part Three. • Files: Discussed in Chapter 12 . • I/O devices: Discussed in Chapter 11 . 3. The OS must protect the data and physical resources of each process against unintended interference by other processes. This involves techniques that relate to memory, files, and I/O devices. A general treatment of protection is found in Chapter 14 . 4. The functioning of a process, and the output it produces, must be independent of the speed at which its execution is carried out relative to the speed of other concurrent processes. This is the subject of this chapter.

10 P I R N O T C E E R S A S C T I O N We can classify the ways in which processes interact on the basis of the degree to which they are aware of each other’s existence. Table 5.2 lists three possible degrees of awareness plus the consequences of each: • Processes unaware of each other: These are independent processes that are not intended to work together. The best example of this situation is the multiprogramming of multiple independent processes. These can either be batch jobs or interactive sessions or a mixture. Although the processes are not working together, the OS needs to be concerned about competition for resources. For example, two independent applications may both want to access the same disk or file or printer. The OS must regulate these accesses. • Processes indirectly aware of each other: These are processes that are not necessarily aware of each other by their respective process IDs but that share access to some object, such as an I/O buffer. Such processes exhibit cooperation in sharing the common object. • Processes directly aware of each other: These are processes that are able to communicate with each other by process ID and that are designed to work jointly on some activity. Again, such processes exhibit cooperation . Conditions will not always be as clear-cut as suggested in Table Rather, several processes may exhibit aspects of both competition and cooperation. Nevertheless, it is productive to examine each of the three items in the preceding list separately and determine their implications for the OS. 10

11 Resource Competition Concurrent processes come into conflict when they use the same resource (competitively or shared) for example: I/O devices, memory, processor time, clock Three control problems must be faced Need for mutual exclusion Deadlock Starvation Sharing processes also need to address coherence Concurrent processes come into conflict with each other when they are competing for the use of the same resource. In its pure form, we can describe the situation as follows. Two or more processes need to access a resource during the course of their execution. Each process is unaware of the existence of other processes, and each is to be unaffected by the execution of the other processes. It follows from this that each process should leave the state of any resource that it uses unaffected. Examples of resources include I/O devices, memory, processor time, and the clock. There is no exchange of information between the competing processes. However, the execution of one process may affect the behavior of competing processes. In particular, if two processes both wish access to a single resource, then one process will be allocated that resource by the OS, and the other will have to wait. Therefore, the process that is denied access will be slowed down. In an extreme case, the blocked process may never get access to the resource and hence will never terminate successfully. In the case of competing processes three control problems must be faced. First is the need for mutual exclusion . Suppose two or more processes require access to a single non-sharable resource, such as a printer. During the course of execution, each process will be sending commands to the I/O device, receiving status information, sending data, and/or receiving data. We will refer to such a resource as a critical resource , and the portion of the program that uses it as a critical section of the program. It is important that only one program at a time be allowed in its critical section. We cannot simply rely on the OS to understand and enforce this restriction because the detailed requirements may not be obvious. In the case of the printer, for example, we want any individual process to have control of the printer while it prints an entire file. Otherwise, lines from competing processes will be interleaved. The enforcement of mutual exclusion creates two additional control problems. One is that of deadlock . For example, consider two processes, P1 and P2, and two resources, R1 and R2. Suppose that each process needs access to both resources to perform part of its function. Then it is possible to have the following situation: the OS assigns R1 to P2, and R2 to P1. Each process is waiting for one of the two resources. Neither will release the resource that it already owns until it has acquired the other resource and performed the function requiring both resources. The two processes are deadlocked. A final control problem is starvation . Suppose that three processes (P1, P2, P3) each require periodic access to resource R. Consider the situation in which P1 is in possession of the resource, and both P2 and P3 are delayed, waiting for that resource. When P1 exits its critical section, either P2 or P3 should be allowed access to R. Assume that the OS grants access to P3 and that P1 again requires access before P3 completes its critical section. If the OS grants access to P1 after P3 has finished, and subsequently alternately grants access to P1 and P3, then P2 may indefinitely be denied access to the resource, even though there is no deadlock situation.

12 Need for Mutual Exclusion
If there is no controlled access to shared data, processes or threads may get an inconsistent view of this data The result of concurrent execution will depend on the order in which instructions are interleaved. Errors are timing dependent and usually not reproducible.

13 A Simple Example Assume P1 and P2 are executing this code and share the variable a Processes can be preempted at any time. Assume P1 is preempted after the input statement, and P2 then executes entirely The character echoed by P1 will be the one read by P2 !! static char a; void echo() { cin >> a; cout << a; }

14 What’s the Problem? This is an example of a race condition Individual processes (threads) execute sequentially in isolation, but concurrency causes them to interact. We need to prevent concurrent execution by processes when they are changing the same data. We need to enforce mutual exclusion.

15 The Critical Section Problem
When a process executes code that manipulates shared data (or resources), we say that the process is in its critical section (CS) for that shared data We must enforce mutual exclusion on the execution of critical sections. Only one process at a time can be in its CS (for that shared data or resource).

16 The Critical Section Problem
Enforcing mutual exclusion guarantees that related CS’s will be executed serially instead of concurrently. The critical section problem is how to provide mechanisms to enforce mutual exclusion so the actions of concurrent processes won’t depend on the order in which their instructions are interleaved

17 The Critical Section Problem
Processes/threads must request permission to enter a CS, & signal when they leave CS. Program structure: entry section: requests entry to CS exit section: notifies that CS is completed remainder section (RS): code that does not involve shared data and resources. The CS problem exists on multiprocessors as well as on uniprocessors.

18 Mutual Exclusion and Data Coherence
Mutual Exclusion ensures data coherence if properly used. Critical Resource (CR) - a shared resource such as a variable, file, or device Data Coherence: The final value or state of a CR shared by concurrently executing processes is the same as the final value or state would be if each process executed serially, in some order.

19 Deadlock and Starvation
Deadlock: two or more processes are blocked permanently because each is waiting for a resource held in a mutually exclusive manner by one of the others. Starvation: a process is repeatedly denied access to some resource which is protected by mutual exclusion, even though the resource periodically becomes available.

20 Mutual Exclusion Control of competition inevitably involves the OS because it is the OS that allocates resources. In addition, the processes themselves will need to be able to express the requirement for mutual exclusion in some fashion, such as locking a resource prior to its use. Any solution will involve some support from the OS, such as the provision of the locking facility. Figure 5.1 illustrates the mutual exclusion mechanism in abstract terms. There are n processes to be executed concurrently. Each process includes (1) a critical section that operates on some resource Ra, and (2) additional code preceding and following the critical section that does not involve access to Ra. Because all processes access the same resource Ra, it is desired that only one process at a time be in its critical section. To enforce mutual exclusion, two functions are provided: entercritical and exitcritical . Each function takes as an argument the name of the resource that is the subject of competition. Any process that attempts to enter its critical section while another process is in its critical section, for the same resource, is made to wait. It remains to examine specific mechanisms for providing the functions entercritical and exitcritical . For the moment, we defer this issue while we consider the other cases of process interaction. Figure Illustration of Mutual Exclusion

21 Requirements for Mutual Exclusion
Mutual Exclusion: must be enforced Non interference: A process that halts must not interfere with other processes No deadlock or starvation Progress : A process must not be denied access to a critical section when there is no other process using it No assumptions are made about relative process speeds or number of processes A process remains inside its critical section for a finite time only Any facility or capability that is to provide support for mutual exclusion should meet the following requirements: 1. Mutual exclusion must be enforced: Only one process at a time is allowed into its critical section, among all processes that have critical sections for the same resource or shared object. 2. A process that halts in its noncritical section must do so without interfering with other processes. 3. It must not be possible for a process requiring access to a critical section to be delayed indefinitely: no deadlock or starvation. 4. When no process is in a critical section, any process that requests entry to its critical section must be permitted to enter without delay. 5. No assumptions are made about relative process speeds or number of processors. 6. A process remains inside its critical section for a finite time only. There are a number of ways in which the requirements for mutual exclusion can be satisfied. One approach is to leave the responsibility with the processes that wish to execute concurrently. Processes, whether they are system programs or application programs, would be required to coordinate with one another to enforce mutual exclusion, with no support from the programming language or the OS. We can refer to these as software approaches. Although this approach is prone to high processing overhead and bugs, it is nevertheless useful to examine such approaches to gain a better understanding of the complexity of concurrent processing. This topic is covered in Appendix A . A second approach involves the use of special purpose machine instructions. These have the advantage of reducing overhead but nevertheless will be shown to be unattractive as a general-purpose solution; they are covered in Section A third approach is to provide some level of support within the OS or a programming language. Three of the most important such approaches are examined in Sections 5.3 through 5.5 .

22 Mutual Exclusion: Hardware Support
Interrupt Disabling uniprocessor system disabling interrupts guarantees mutual exclusion Disadvantages: the efficiency of execution could be noticeably degraded this approach will not work in a multiprocessor architecture

23 Mutual Exclusion: Hardware Support
Special Machine Instructions Compare&Swap Instruction also called a “compare and exchange instruction” a compare is made between a memory value and a test value if the old memory value = test value, swap in a new value to the memory location always return the old memory value carried out atomically in the hardware. COMPARE&SWAP INSTRUCTION This version of the instruction checks a memory location ( *word ) against a test value ( testval ). If the memory location’s current value is testval, it is replaced with newval ; otherwise it is left unchanged. The old memory value is always returned; thus, the memory location has been updated if the returned value is the same as the test value. This atomic instruction therefore has two parts: A compare is made between a memory value and a test value; if the values are the same, a swap occurs. The entire compare&swap function is carried out atomically—that is, it is not subject to interruption. Another version of this instruction returns a Boolean value: true if the swap occurred; false otherwise. Some version of this instruction is available on nearly all processor families (x86, IA64, sparc, /390, etc.), and most operating systems use this instruction for support of concurrency.

24 MutualExclusion:Hardware Support
Compare&Swap Instruction Pseudo-code definition of the hardware instruction: int compare_and_swap (word, test_val, new_val) { Int oldval; Oldval = word if (word ==test_val) word = new_val; return oldval COMPARE&SWAP INSTRUCTION This version of the instruction checks a memory location ( *word ) against a test value ( testval ). If the memory location’s current value is testval, it is replaced with newval ; otherwise it is left unchanged. The old memory value is always returned; thus, the memory location has been updated if the returned value is the same as the test value. This atomic instruction therefore has two parts: A compare is made between a memory value and a test value; if the values are the same, a swap occurs. The entire compare&swap function is carried out atomically—that is, it is not subject to interruption. Another version of this instruction returns a Boolean value: true if the swap occurred; false otherwise. Some version of this instruction is available on nearly all processor families (x86, IA64, sparc, /390, etc.), and most operating systems use this instruction for support of concurrency.

25 Compare and Swap Instruction
word = bolt test_val = 0 new_val = 1 If bolt is 0 when the C&S is executed, the condition is false and P enters its critical section. (leaves bolt = 1) If bolt = 1 when C&S executes, P continues to execute the while loop. It’s busy waiting ( or spinning) Figure 5.2a shows a mutual exclusion protocol based on the use of this instruction. A shared variable bolt is initialized to 0. The only process that may enter its critical section is one that finds bolt equal to 0. All other processes attempting to enter their critical section go into a busy waiting mode. The term busy waiting , or spin waiting , refers to a technique in which a process can do nothing until it gets permission to enter its critical section but continues to execute an instruction or set of instructions that tests the appropriate variable to gain entrance. When a process leaves its critical section, it resets bolt to 0; at this point one and only one of the waiting processes is granted access to its critical section. The choice of process depends on which process happens to execute the compare & swap instruction next. Figure 5.2 Hardware Support for Mutual Exclusion

26 Exchange Instruction Figure 5.2 Hardware Support for Mutual Exclusion
Figure 5.2b shows a mutual exclusion protocol based on the use of an exchange instruction. A shared variable bolt is initialized to 0. Each process uses a local variable key that is initialized to 1. The only process that may enter its critical section is one that finds bolt equal to 0. It excludes all other processes from the critical section by setting bolt to 1. When a process leaves its critical section, it resets bolt to 0, allowing another process to gain access to its critical section. Figure 5.2 Hardware Support for Mutual Exclusion

27 Special Machine Instruction: Advantages
Applicable to any number of processes on either a single processor or multiple processors sharing main memory Simple and easy to verify It can be used to support multiple critical sections; each critical section can be defined by its own variable The use of a special machine instruction to enforce mutual exclusion has a number of advantages: • It is applicable to any number of processes on either a single processor or multiple processors sharing main memory. • It is simple and therefore easy to verify. • It can be used to support multiple critical sections; each critical section can be defined by its own variable.

28 Special Machine Instruction: Disadvantages
Busy-waiting is employed, thus while a process is waiting for access to a critical section it continues to consume processor time Starvation is possible when a process leaves a critical section and more than one process is waiting Deadlock is possible if priority- based scheduling is used There are some serious disadvantages: • Busy waiting is employed: Thus, while a process is waiting for access to a critical section, it continues to consume processor time. Starvation is possible: When a process leaves a critical section and more than one process is waiting, the selection of a waiting process is arbitrary. Thus, some process could indefinitely be denied access. • Deadlock is possible: Consider the following scenario on a single-processor system. Process P1 executes the special instruction (e.g., compare & swap, exchange ) and enters its critical section. P1 is then interrupted to give the processor to P2, which has higher priority. If P2 now attempts to use the same resource as P1, it will be denied access because of the mutual exclusion mechanism. Thus, it will go into a busy waiting loop. However, P1 will never be dispatched because it is of lower priority than another ready process, P2. Because of the drawbacks of both the software and hardware solutions just outlined, we need to look for other mechanisms.

29 Common Concurrency Mechanisms
We now turn to OS and programming language mechanisms that are used to provide concurrency. Table 5.3 summarizes mechanisms in common use. We begin, in this section, with semaphores. The next two sections discuss monitors and message passing. The other mechanisms in Table 5.3 are discussed when treating specific operating system examples, in Chapters 6 and 13 .

30 Semaphore There is no way to inspect or manipulate semaphores other than these three operations A variable that has an integer value upon which only three operations are defined: The fundamental principle is this: Two or more processes can cooperate by means of simple signals, such that a process can be forced to stop at a specified place until it has received a specific signal. Any complex coordination requirement can be satisfied by the appropriate structure of signals. For signaling, special variables called semaphores are used. To transmit a signal via semaphore s , a process executes the primitive semSignal(s) . To receive a signal via semaphore s , a process executes the primitive semWait(s) ; if the corresponding signal has not yet been transmitted, the process is suspended until the transmission takes place. To achieve the desired effect, we can view the semaphore as a variable that has an integer value upon which only three operations are defined: 1. A semaphore may be initialized to a nonnegative integer value. 2. The semWait operation decrements the semaphore value. If the value becomes negative, then the process executing the semWait is blocked. Otherwise, the process continues execution. 3. The semSignal operation increments the semaphore value. If the resulting value is less than or equal to zero, then a process blocked by a semWait operation, if any, is unblocked. Other than these three operations, there is no way to inspect or manipulate semaphores. We explain these operations as follows. To begin, the semaphore has a zero or positive value. If the value is positive, that value equals the number of processes that can issue a wait and immediately continue to execute. If the value is zero, either by initialization or because a number of processes equal to the initial semaphore value have issued a wait, the next process to issue a wait is blocked, and the semaphore value goes negative. Each subsequent wait drives the semaphore value further into minus territory. The negative value equals the number of processes waiting to be unblocked. Each signal unblocks one of the waiting processes when the semaphore value is negative. May be initialized to a nonnegative integer value The semWait operation decrements the value The semSignal operation increments the value

31 Consequences There is no way to know before a process decrements a semaphore whether it will block or not There is no way to know which process will continue immediately on a uniprocessor system when two processes are running concurrently You don’t know whether another process is waiting so the number of unblocked processes may be zero or one [DOWN08] points out three interesting consequences of the semaphore definition: • In general, there is no way to know before a process decrements a semaphore whether it will block or not. After a process increments a semaphore and another process gets woken up, both processes continue running concurrently. There is no way to know which process, if either, will continue immediately on a uniprocessor system. • When you signal a semaphore, you don’t necessarily know whether another process is waiting, so the number of unblocked processes may be zero or one.

32 Semaphore Primitives Figure 5.3 suggests a more formal definition of the primitives for semaphores. The semWait and semSignal primitives are assumed to be atomic.

33 Binary Semaphore Primitives
A more restricted version, known as the binary semaphore , is defined in Figure A binary semaphore may only take on the values 0 and 1 and can be defined by the following three operations: 1. A binary semaphore may be initialized to 0 or 1. 2. The semWaitB operation checks the semaphore value. If the value is zero, then the process executing the semWaitB is blocked. If the value is one, then the value is changed to zero and the process continues execution. 3. The semSignalB operation checks to see if any processes are blocked on this semaphore (semaphore value equals 0). If so, then a process blocked by a semWaitB operation is unblocked. If no processes are blocked, then the value of the semaphore is set to one. In principle, it should be easier to implement the binary semaphore, and it can be shown that it has the same expressive power as the general semaphore (see Problem 5.17). To contrast the two types of semaphores, the nonbinary semaphore is often referred to as either a counting semaphore or a general semaphore . A concept related to the binary semaphore is the mutex . A key difference between the two is that the process that locks the mutex (sets the value to zero) must be the one to unlock it (sets the value to 1). In contrast, it is possible for one process to lock a binary semaphore and for another to unlock it.

34 Strong/Weak Semaphores
A queue is used to hold processes waiting on the semaphore the process that has been blocked the longest is released from the queue first (FIFO) Strong Semaphores the order in which processes are removed from the queue is not specified Weak Semaphores For both counting semaphores and binary semaphores, a queue is used to hold processes waiting on the semaphore. The question arises of the order in which processes are removed from such a queue. The fairest removal policy is first-in-first-out (FIFO): The process that has been blocked the longest is released from the queue first; a semaphore whose definition includes this policy is called a strong semaphore . A semaphore that does not specify the order in which processes are removed from the queue is a weak semaphore .

35 Example of Semaphore Mechanism
Figure 5.5 , based on one in [DENN84], is an example of the operation of a strong semaphore. Here processes A, B, and C depend on a result from process D. Initially (1), A is running; B, C, and D are ready; and the semaphore count is 1, indicating that one of D’s results is available. When A issues a semWait instruction on semaphore s , the semaphore decrements to 0, and A can continue to execute; subsequently it rejoins the ready queue. Then B runs (2), eventually issues a semWait instruction, and is blocked, allowing D to run (3). When D completes a new result, it issues a semSignal instruction, which allows B to move to the ready queue (4). D rejoins the ready queue and C begins to run (5) but is blocked when it issues a semWait instruction. Similarly, A and B run and are blocked on the semaphore, allowing D to resume execution (6). When D has a result, it issues a semSignal , which transfers C to the ready queue. Later cycles of D will release A and B from the Blocked state.

36 Mutual Exclusion For the mutual exclusion algorithm discussed in the next subsection and illustrated in Figure 5.6 , strong semaphores guarantee freedom from starvation, while weak semaphores do not. We will assume strong semaphores because they are more convenient and because this is the form of semaphore typically provided by operating systems. Figure 5.6 shows a straightforward solution to the mutual exclusion problem using a semaphore s (compare Figure 5.1 ). Consider n processes, identified in the array P ( i ), all of which need access to the same resource. Each process has a critical section used to access the resource. In each process, a semWait(s) is executed just before its critical section. If the value of s becomes negative, the process is blocked. If the value is 1, then it is decremented to 0 and the process immediately enters its critical section; because s is no longer positive, no other process will be able to enter its critical section. The semaphore is initialized to 1. Thus, the first process that executes a semWait will be able to enter the critical section immediately, setting the value of s to 0. Any other process attempting to enter the critical section will find it busy and will be blocked, setting the value of s to –1. Any number of processes may attempt entry; each such unsuccessful attempt results in a further decrement of the value of s . When the process that initially entered its critical section departs, s is incremented and one of the blocked processes (if any) is removed from the queue of blocked processes associated with the semaphore and put in a Ready state. When it is next scheduled by the OS, it may enter the critical section.

37 Shared Data Protected by a Semaphore
Figure 5.7 , based on one in [BACO03], shows a possible sequence for three processes using the mutual exclusion discipline of Figure In this example three processes (A, B, C) access a shared resource protected by the semaphore lock . Process A executes semWait( lock ) ; because the semaphore has a value of 1 at the time of the semWait operation, A can immediately enter its critical section and the semaphore takes on the value 0. While A is in its critical section, both B and CFigure 5.7 , based on one in [BACO03], shows a possible sequence for three processes using the mutual exclusion discipline of Figure In this example three processes (A, B, C) access a shared resource protected by the semaphore lock . Process A executes semWait( lock ) ; because the semaphore has a value of 1 at the time of the semWait operation, A can immediately enter its critical section and the semaphore takes on the value 0. While A is in its critical section, both B and C perform a semWait operation and are blocked pending the availability of the semaphore. When A exits its critical section and performs semSignal( lock ) , B, which was the first process in the queue, can now enter its critical section.

38 Producer/Consumer Problem
General Situation: one or more producers are generating data and placing these in a buffer a single consumer is taking items out of the buffer one at time only one producer or consumer may access the buffer at any one time The Problem: ensure that the producer can’t add data into full buffer and consumer can’t remove data from an empty buffer To begin, let us assume that the buffer is infinite and consists of a linear array of elements. We now examine one of the most common problems faced in concurrent processing: the producer/consumer problem. The general statement is this: There are one or more producers generating some type of data (records, characters) and placing these in a buffer. There is a single consumer that is taking items out of the buffer one at a time. The system is to be constrained to prevent the overlap of buffer operations. That is, only one agent (producer or consumer) may access the buffer at any one time. The problem is to make sure that the producer won’t try to add data into the buffer if it’s full and that the consumer won’t try to remove data from an empty buffer. We will look at a number of solutions to this problem to illustrate both the power and the pitfalls of semaphores.

39 Buffer Structure Figure 5.8 illustrates the structure of buffer b . The producer can generate items and store them in the buffer at its own pace. Each time, an index ( in ) into the buffer is incremented. The consumer proceeds in a similar fashion but must make sure that it does not attempt to read from an empty buffer. Hence, the consumer makes sure that the producer has advanced beyond it ( in > out ) before proceeding.

40 Incorrect Solution Let us try to implement this system using binary semaphores. Figure 5.9 is a first attempt. Rather than deal with the indices in and out , we can simply keep track of the number of items in the buffer, using the integer variable n (= in – out ). The semaphore s is used to enforce mutual exclusion; the semaphore delay is used to force the consumer to semWait if the buffer is empty. This solution seems rather straightforward. The producer is free to add to the buffer at any time. It performs semWaitB(s) before appending and semSignalB(s) afterward to prevent the consumer or any other producer from accessing the buffer during the append operation. Also, while in the critical section, the producer increments the value of n . If n = 1, then the buffer was empty just prior to this append, so the producer performs semSignalB(delay) to alert the consumer of this fact. The consumer begins by waiting for the first item to be produced, using semWaitB(delay) . It then takes an item and decrements n in its critical section. If the producer is able to stay ahead of the consumer (a common situation), then the consumer will rarely block on the semaphore delay because n will usually be positive. Hence both producer and consumer run smoothly. There is, however, a flaw in this program. When the consumer has exhausted the buffer, it needs to reset the delay semaphore so that it will be forced to wait until the producer has placed more items in the buffer. This is the purpose of the statement: if n == 0 semWaitB(delay) . Figure An Incorrect Solution to the Infinite-Buffer Producer/Consumer Problem Using Binary Semaphores

41 Possible Solution Consider the scenario outlined in Table In line 14, the consumer fails to execute the semWaitB operation. The consumer did indeed exhaust the buffer and set n to 0 (line 8), but the producer has incremented n before the consumer can test it in line 14. The result is a semSignalB not matched by a prior semWaitB . The value of –1 for n in line 20 means that the consumer has consumed an item from the buffer that does not exist. It would not do simply to move the conditional statement inside the critical section of the consumer because this could lead to deadlock (e.g., after line 8 of Table 5.4 ).

42 Solution Correct A fix for the problem is to introduce an auxiliary variable that can be set in the consumer’s critical section for use later on. This is shown in Figure A careful trace of the logic should convince you that deadlock can no longer occur. Figure A Correct Solution to the Infinite-Buffer Producer/Consumer Problem Using Binary Semaphores

43 Solution Using Semaphores
A somewhat cleaner solution can be obtained if general semaphores (also called counting semaphores) are used, as shown in Figure The variable n is now a semaphore. Its value still is equal to the number of items in the buffer. Suppose now that in transcribing this program, a mistake is made and the operations semSignal(s) and semSignal(n) are interchanged. This would require that the semSignal(n) operation be performed in the producer’s critical section without interruption by the consumer or another producer. Would this affect the program? No, because the consumer must wait on both semaphores before proceeding in any case.

44 Finite Circular Buffer
Finally, let us add a new and realistic restriction to the producer/consumer problem: namely, that the buffer is finite. The buffer is treated as a circular storage ( Figure 5.12 ), and pointer values must be expressed modulo the size of the buffer.

45 Solution Using Semaphores
Figure 5.13 shows a solution using general semaphores. The semaphore e has been added to keep track of the number of empty spaces. Figure A Solution to the Bounded-Buffer Producer/Consumer Problem Using Semaphores

46 Implementation of Semaphores
Imperative that the semWait and semSignal operations be implemented as atomic primitives Can be implemented in hardware or firmware Software schemes such as Dekker’s or Peterson’s algorithms can be used Use one of the hardware-supported schemes for mutual exclusion As was mentioned earlier, it is imperative that the semWait and semSignal operations be implemented as atomic primitives. One obvious way is to implement them in hardware or firmware. Failing this, a variety of schemes have been suggested. The essence of the problem is one of mutual exclusion: Only one process at a time may manipulate a semaphore with either a semWait or semSignal operation. Thus, any of the software schemes, such as Dekker’s algorithm or Peterson’s algorithm ( Appendix A ), could be used; this would entail a substantial processing overhead. Another alternative is to use one of the hardware-supported schemes for mutual exclusion.

47 Review Concurrent processes, threads Access to shared data/resources
Need to enforce mutual exclusion Hardware mechanisms have limited usefulness Semaphores: OS mechanism for mutual exclusion & other synchronization issues Standard semaphore/counting Binary semaphore Producer/consumer problem

48 Implemented in a number of programming languages
Monitors Programming language construct that provides equivalent functionality to that of semaphores and is easier to control Implemented in a number of programming languages including Concurrent Pascal, Pascal-Plus, Modula-2, Modula-3, and Java Has also been implemented as a program library Software module consisting of one or more procedures, an initialization sequence, and local data The monitor is a programming-language construct that provides equivalent functionality to that of semaphores and that is easier to control. The concept was first formally defined in [HOAR74]. The monitor construct has been implemented in a number of programming languages, including Concurrent Pascal, Pascal-Plus, Modula-2, Modula-3, and Java. It has also been implemented as a program library. This allows programmers to put a monitor lock on any object. In particular, for something like a linked list, you may want to lock all linked lists with one lock, or have one lock for each list, or have one lock for each element of each list. A monitor is a software module consisting of one or more procedures, an initialization sequence, and local data.

49 Monitor Characteristics
Local data variables are accessible only by the monitor’s procedures and not by any external procedure Process enters monitor by invoking one of its procedures Only one process may be executing in the monitor at a time The chief characteristics of a monitor are the following: 1. The local data variables are accessible only by the monitor’s procedures and not by any external procedure. 2. A process enters the monitor by invoking one of its procedures. 3. Only one process may be executing in the monitor at a time; any other processes that have invoked the monitor are blocked, waiting for the monitor to become available. The first two characteristics are reminiscent of those for objects in object-oriented software. Indeed, an object-oriented OS or programming language can readily implement a monitor as an object with special characteristics. By enforcing the discipline of one process at a time, the monitor is able to provide a mutual exclusion facility. The data variables in the monitor can be accessed by only one process at a time. Thus, a shared data structure can be protected by placing it in a monitor. If the data in a monitor represent some resource, then the monitor provides a mutual exclusion facility for accessing the resource. To be useful for concurrent processing, the monitor must include synchronization tools. For example, suppose a process invokes the monitor and, while in the monitor, must be blocked until some condition is satisfied. A facility is needed by which the process is not only blocked but releases the monitor so that some other process may enter it. Later, when the condition is satisfied and the monitor is again available, the process needs to be resumed and allowed to reenter the monitor at the point of its suspension.

50 Synchronization Achieved by the use of condition variables that are contained within the monitor and accessible only within the monitor Condition variables are operated on by two functions: cwait(c): suspend execution of the calling process on condition c csignal(c): resume execution of some process blocked after a cwait on the same condition A monitor supports synchronization by the use of condition variables that are contained within the monitor and accessible only within the monitor. Condition variables are a special data type in monitors, which are operated on by two functions: • cwait(c) : Suspend execution of the calling process on condition c . The monitor is now available for use by another process. • csignal(c) : Resume execution of some process blocked after a cwait on the same condition. If there are several such processes, choose one of them; if there is no such process, do nothing. Note that monitor wait and signal operations are different from those for the semaphore. If a process in a monitor signals and no task is waiting on the condition variable, the signal is lost.

51 Structure of a Monitor Figure 5.15 Structure of a Monitor
Figure 5.15 illustrates the structure of a monitor. Although a process can enter the monitor by invoking any of its procedures, we can think of the monitor as having a single entry point that is guarded so that only one process may be in the monitor at a time. Other processes that attempt to enter the monitor join a queue of processes blocked waiting for monitor availability. Once a process is in the monitor, it may temporarily block itself on condition x by issuing cwait(x) ; it is then placed in a queue of processes waiting to reenter the monitor when the condition changes, and resume execution at the point in its program following the cwait(x) call. If a process that is executing in the monitor detects a change in the condition variable x , it issues csignal(x) , which alerts the corresponding condition queue that the condition has changed. Figure Structure of a Monitor

52 Problem Solution Using a Monitor
As an example of the use of a monitor, let us return to the bounded-buffer producer/consumer problem. Figure 5.16 shows a solution using a monitor. The monitor module, boundedbuffer , controls the buffer used to store and retrieve characters. The monitor includes two condition variables (declared with the construct cond ): notfull is true when there is room to add at least one character to the buffer, and notempty is true when there is at least one character in the buffer. A producer can add characters to the buffer only by means of the procedure append inside the monitor; the producer does not have direct access to buffer . The procedure first checks the condition notfull to determine if there is space available in the buffer. If not, the process executing the monitor is blocked on that condition. Some other process (producer or consumer) may now enter the monitor. Later, when the buffer is no longer full, the blocked process may be removed from the queue, reactivated, and resume processing. After placing a character in the buffer, the process signals the notempty condition. A similar description can be made of the consumer function. This example points out the division of responsibility with monitors compared to semaphores. In the case of monitors, the monitor construct itself enforces mutual exclusion: It is not possible for both a producer and a consumer simultaneously to access the buffer. However, the programmer must place the appropriate cwait and csignal primitives inside the monitor to prevent processes from depositing items in a full buffer or removing them from an empty one. In the case of semaphores, both mutual exclusion and synchronization are the responsibility of the programmer. Note that in Figure 5.16 , a process exits the monitor immediately after executing the csignal function. If the csignal does not occur at the end of the procedure, then, in Hoare’s proposal, the process issuing the signal is blocked to make the monitor available and placed in a queue until the monitor is free. One possibility at this point would be to place the blocked process in the entrance queue, so that it would have to compete for access with other processes that had not yet entered the monitor. However, because a process blocked on a csignal function has already partially performed its task in the monitor, it makes sense to give this process precedence over newly entering processes by setting up a separate urgent queue ( Figure 5.15 ). One language that uses monitors, Concurrent Pascal, requires that csignal only appear as the last operation executed by a monitor procedure. If there are no processes waiting on condition x , then the execution of csignal( x ) has no effect. Figure A Solution to the Bounded-Buffer Producer/Consumer Problem Using a Monitor

53 Readers/Writers Problem
A data area is shared among many processes some processes only read the data area, (readers) and some only write to the data area (writers) Conditions that must be satisfied: any number of readers may simultaneously read the file only one writer at a time may write to the file if a writer is writing to the file, no reader may read it The readers/writers problem is defined as follows: There is a data area shared among a number of processes. The data area could be a file, a block of main memory, or even a bank of processor registers. There are a number of processes that only read the data area (readers) and a number that only write to the data area (writers). The conditions that must be satisfied are as follows: 1. Any number of readers may simultaneously read the file. 2. Only one writer at a time may write to the file. 3. If a writer is writing to the file, no reader may read it. Thus, readers are processes that are not required to exclude one another and writers are processes that are required to exclude all other processes, readers and writers alike.

54 Readers Have Priority Solution
Figure 5.22 is a solution using semaphores, showing one instance each of a reader and a writer; the solution does not change for multiple readers and writers. The writer process is simple. The semaphore wsem is used to enforce mutual exclusion. As long as one writer is accessing the shared data area, no other writers and no readers may access it. The reader process also makes use of wsem to enforce mutual exclusion. However, to allow multiple readers, we require that, when there are no readers reading, the first reader that attempts to read should wait on wsem . When there is already at least one reader reading, subsequent readers need not wait before entering. The global variable readcount is used to keep track of the number of readers, and the semaphore x is used to assure that readcount is updated properly. Figure A Solution to the Readers/Writers Problem Using Semaphore: Readers Have Priority

55 Solution: Writers Have Priority
In the previous solution, readers have priority. Once a single reader has begun to access the data area, it is possible for readers to retain control of the data area as long as there is at least one reader in the act of reading. Therefore, writers are subject to starvation. Figure 5.23 shows a solution that guarantees that no new readers are allowed access to the data area once at least one writer has declared a desire to write. For writers, the following semaphores and variables are added to the ones already defined: • A semaphore rsem that inhibits all readers while there is at least one writer desiring access to the data area • A variable writecount that controls the setting of rsem • A semaphore y that controls the updating of writecount Figure A Solution to the Readers/Writers Problem Using Semaphore: Writers Have Priority

56 State of the Process Queues
For readers, one additional semaphore is needed. A long queue must not be allowed to build up on rsem ; otherwise writers will not be able to jump the queue. Therefore, only one reader is allowed to queue on rsem , with any additional readers queuing on semaphore z , immediately before waiting on rsem . Table 5.6 summarizes the possibilities.

57 Message Passing An alternative solution, which gives writers priority and which is implemented using message passing, is shown in Figure In this case, there is a controller process that has access to the shared data area. Other processes wishing to access the data area send a request message to the controller, are granted access with an “OK” reply message, and indicate completion of access with a “finished” message. The controller is equipped with three mailboxes, one for each type of message that it may receive. The controller process services write request messages before read request messages to give writers priority. In addition, mutual exclusion must be enforced. To do this the variable count is used, which is initialized to some number greater than the maximum possible number of readers. In this example, we use a value of 100. The action of the controller can be summarized as follows: • If count > 0, then no writer is waiting and there may or may not be readers active. Service all “finished” messages first to clear active readers. Then service write requests and then read requests. • If count = 0, then the only request outstanding is a write request. Allow the writer to proceed and wait for a “finished” message. If count < 0, then a writer has made a request and is being made to wait to clear all active readers. Therefore, only “finished” messages should be serviced. Figure A Solution to the Readers/Writers Problem Using Message Passing

58 Deadlock The permanent blocking of a set of processes that either compete for system resources or communicate with each other A set of processes is deadlocked when each process in the set is blocked awaiting an event that can only be triggered by another blocked process in the set Permanent No efficient solution Deadlock can be defined as the permanent blocking of a set of processes that either compete for system resources or communicate with each other. A set of processes is deadlocked when each process in the set is blocked awaiting an event (typically the freeing up of some requested resource) that can only be triggered by another blocked process in the set. Deadlock is permanent because none of the events is ever triggered. Unlike other problems in concurrent process management, there is no efficient solution in the general case.

59 Potential Deadlock I need quad C and D I need quad B and C
I need quad A and B Animated Slide Click 1 Cars approach intersection Then Cars announce their resource needs All deadlocks involve conflicting needs for resources by two or more processes. A common example is the traffic deadlock. Figure 6.1a shows a situation in which four cars have arrived at a four-way stop intersection at approximately the same time. The four quadrants of the intersection are the resources over which control is needed. In particular, if all four cars wish to go straight through the intersection, the resource requirements are as follows: • Car 1, traveling north, needs quadrants a and b. • Car 2 needs quadrants b and c. • Car 3 needs quadrants c and d. • Car 4 needs quadrants d and a. The typical rule of the road in the United States is that a car at a four-way stop should defer to a car immediately to its right. This rule works if there are only two or three cars at the intersection. For example, if only the northbound and westbound cars arrive at the intersection, the northbound car will wait and the westbound car proceeds. However, if all four cars arrive at about the same time, each will refrain from entering the intersection, this causes a potential deadlock. The deadlock is only potential, not actual, because the necessary resources are available for any of the cars to proceed. If one car eventually does proceed, it can do so. I need quad D and A

60 Actual Deadlock HALT until D is free HALT until C is free
HALT until B is free Animated Slide Click 1 Cars move to deadlock Then Cars announce their resource need However, if all four cars ignore the rules and proceed (cautiously) into the intersection at the same time, then each car seizes one resource (one quadrant) but cannot proceed because the required second resource has already been seized by another car. This is an actual deadlock. HALT until A is free

61 Joint Progress Diagram
Let us now look at a depiction of deadlock involving processes and computer resources. Figure 6.2 (based on one in [BACO03]), which we refer to as a joint progress diagram , illustrates the progress of two processes competing for two resources. Each process needs exclusive use of both resources for a certain period of time. In Figure 6.2 , the x -axis represents progress in the execution of P and the y –axis represents progress in the execution of Q. The joint progress of the two processes is therefore represented by a path that progresses from the origin in a northeasterly direction. For a uniprocessor system, only one process at a time may execute, and the path consists of alternating horizontal and vertical segments, with a horizontal segment representing a period when P executes and Q waits and a vertical segment representing a period when Q executes and P waits. The figure indicates areas in which both P and Q require resource A (upward slanted lines); both P and Q require resource B (downward slanted lines); and both P and Q require both resources. Because we assume that each process requires exclusive control of any resource, these are all forbidden regions; that is, it is impossible for any path representing the joint execution progress of P and Q to enter these regions. The figure shows six different execution paths. These can be summarized as follows: 1. Q acquires B and then A and then releases B and A. When P resumes execution, it will be able to acquire both resources. 2. Q acquires B and then A. P executes and blocks on a request for A. Q releases B and A. When P resumes execution, it will be able to acquire both resources. 3. Q acquires B and then P acquires A. Deadlock is inevitable, because as execution proceeds, Q will block on A and P will block on B. 4. P acquires A and then Q acquires B. Deadlock is inevitable, because as execution 5. P acquires A and then B. Q executes and blocks on a request for B. P releases A and B. When Q resumes execution, it will be able to acquire both resources. 6. P acquires A and then B and then releases A and B. When Q resumes execution, The gray-shaded area of Figure 6.2 , which can be referred to as a fatal region , applies to the commentary on paths 3 and 4. If an execution path enters this fatal region, then deadlock is inevitable. Note that the existence of a fatal region depends on the logic of the two processes. However, deadlock is only inevitable if the joint progress of the two processes creates a path that enters the fatal region.

62 No Deadlock Example Whether or not deadlock occurs depends on both the dynamics of the execution and on the details of the application. This situation is reflected in Figure Some thought should convince you that regardless of the relative timing of the two processes, deadlock cannot occur. As shown, the joint progress diagram can be used to record the execution history of two processes that share resources. In cases where more than two processes may compete for the same resource, a higher-dimensional diagram would be required. The principles concerning fatal regions and deadlock would remain the same.

63 Resource Categories Reusable Consumable
can be safely used by only one process at a time and is not depleted by that use processors, I/O channels, main and secondary memory, devices, and data structures such as files, databases, and semaphores Consumable one that can be created (produced) and destroyed (consumed) interrupts, signals, messages, and information in I/O buffers Two general categories of resources can be distinguished: reusable and consumable. A reusable resource is one that can be safely used by only one process at a time and is not depleted by that use. Processes obtain resource units that they later release for reuse by other processes. Examples of reusable resources include processors; I/O channels; main and secondary memory; devices; and data structures such as files, databases, and semaphores. A consumable resource is one that can be created (produced) and destroyed (consumed). Examples of consumable resources are interrupts, signals, messages, and information in I/O buffers.

64 Reusable Resources Example
As an example of deadlock involving reusable resources, consider two processes that compete for exclusive access to a disk file D and a tape drive T. The programs engage in the operations depicted in Figure Deadlock occurs if each process holds one resource and requests the other. For example, deadlock occurs if the multiprogramming system interleaves the execution of the two processes as follows: p 0 p 1 q 0 q 1 p 2 q 2 It may appear that this is a programming error rather than a problem for the OS designer. However, we have seen that concurrent program design is challenging. Such deadlocks do occur, and the cause is often embedded in complex program logic, making detection difficult. One strategy for dealing with such a deadlock is to impose system design constraints concerning the order in which resources can be requested.

65 Example 2: Memory Request
Space is available for allocation of 200Kbytes, and the following sequence of events occur: Deadlock occurs if both processes progress to their second request P1 P2 . . . . . . Request 80 Kbytes; Request 70 Kbytes; . . . . . . Request 60 Kbytes; Request 80 Kbytes; Another example of deadlock with a reusable resource has to do with requests for main memory. Suppose the space available for allocation is 200 Kbytes, and the following sequence of requests occurs: (see diagram) Deadlock occurs if both processes progress to their second request. If the amount of memory to be requested is not known ahead of time, it is difficult to deal with this type of deadlock by means of system design constraints. The best way to deal with this particular problem is, in effect, to eliminate the possibility by using virtual memory, which is discussed in Chapter 8 .

66 Consumable Resources Deadlock
Consider a pair of processes, in which each process attempts to receive a message from the other process and then send a message to the other process: Deadlock occurs if the Receive is blocking As an example of deadlock involving consumable resources, consider the following pair of processes, in which each process attempts to receive a message from the other process and then send a message to the other process: (see diagram) Deadlock occurs if the Receive is blocking (i.e., the receiving process is blocked until the message is received). Once again, a design error is the cause of the deadlock. Such errors may be quite subtle and difficult to detect. Furthermore, it may take a rare combination of events to cause the deadlock; thus a program could be in use for a considerable period of time, even years, before the deadlock actually occurs.

67 Deadlock Detection, Prevention, and Avoidance
There is no single effective strategy that can deal with all types of deadlock. Table 6.1 summarizes the key elements of the most important approaches that have been developed: prevention, avoidance, and detection. We examine each of these in turn, after first introducing resource allocation graphs and then discussing the conditions for deadlock.

68 Resource Allocation Graphs
A useful tool in characterizing the allocation of resources to processes is the resource allocation graph , introduced by Holt [HOLT72]. The resource allocation graph is a directed graph that depicts a state of the system of resources and processes, with each process and each resource represented by a node. A graph edge directed from a process to a resource indicates a resource that has been requested by the process but not yet granted ( Figure 6.5a ). Within a resource node, a dot is shown for each instance of that resource. Examples of resource types that may have multiple instances are I/O devices that are allocated by a resource management module in the OS. A graph edge directed from a reusable resource node dot to a process indicates a request that has been granted ( Figure 6.5b ); that is, the process has been assigned one unit of that resource. A graph edge directed from a consumable resource node dot to a process indicates that the process is the producer of that resource. Figure 6.5c shows an example deadlock. There is only one unit each of resources Ra and Rb. Process P1 holds Rb and requests Ra, while P2 holds Ra but requests Rb. Figure 6.5d has the same topology as Figure 6.5c , but there is no deadlock because multiple units of each resource are available.

69 Resource Allocation Graphs
The resource allocation graph of Figure 6.6 corresponds to the deadlock situation in Figure 6.1b . Note that in this case, we do not have a simple situation in which two processes each have one resource the other needs. Rather, in this case, there is a circular chain of processes and resources that results in deadlock.

70 Conditions for Deadlock
Mutual Exclusion only one process may use a resource at a time Hold-and-Wait a process may hold allocated resources while awaiting assignment of others No Pre-emption no resource can be forcibly removed from a process holding it Circular Wait a closed chain of processes exists, such that each process holds at least one resource needed by the next process in the chain Three conditions of policy must be present for a deadlock to be possible: 1. Mutual exclusion . Only one process may use a resource at a time. No process may access a resource unit that has been allocated to another process. 2. Hold and wait . A process may hold allocated resources while awaiting assignment of other resources. 3. No preemption . No resource can be forcibly removed from a process holding it. In many ways these conditions are quite desirable. For example, mutual exclusion is needed to ensure consistency of results and the integrity of a database. Similarly, preemption should not be done arbitrarily. For example, when data resources are involved, preemption must be supported by a rollback recovery mechanism, which restores a process and its resources to a suitable previous state from which the process can eventually repeat its actions. The first three conditions are necessary but not sufficient for a deadlock to exist. For deadlock to actually take place, a fourth condition is required: 4. Circular wait . A closed chain of processes exists, such that each process holds at least one resource needed by the next process in the chain (e.g., Figure 6.5c and Figure 6.6 ). The fourth condition is, actually, a potential consequence of the first three. That is, given that the first three conditions exist, a sequence of events may occur that lead to an unresolvable circular wait. The unresolvable circular wait is in fact the definition of deadlock. The circular wait listed as condition 4 is unresolvable because the first three conditions hold. Thus, the four conditions, taken together, constitute necessary and sufficient conditions for deadlock.

71 Dealing with Deadlock Prevent Deadlock
Three general approaches exist for dealing with deadlock: adopt a policy that eliminates one of the conditions Prevent Deadlock make the appropriate dynamic choices based on the current state of resource allocation Avoid Deadlock attempt to detect the presence of deadlock and take action to recover Detect Deadlock Three general approaches exist for dealing with deadlock. First, one can prevent deadlock by adopting a policy that eliminates one of the conditions (conditions 1 through 4). Second, one can avoid deadlock by making the appropriate dynamic choices based on the current state of resource allocation. Third, one can attempt to detect the presence of deadlock (conditions 1 through 4 hold) and take action to recover.

72 Deadlock Prevention Strategy
Design a system in such a way that the possibility of deadlock is excluded Two main methods: Indirect prevent the occurrence of one of the three necessary conditions Direct prevent the occurrence of a circular wait The strategy of deadlock prevention is, simply put, to design a system in such a way that the possibility of deadlock is excluded. We can view deadlock prevention methods as falling into two classes. An indirect method of deadlock prevention is to prevent the occurrence of one of the three necessary conditions listed previously (items 1 through 3). A direct method of deadlock prevention is to prevent the occurrence of a circular wait (item 4).

73 Deadlock Condition Prevention
Mutual Exclusion if access to a resource requires mutual exclusion then it must be supported by the OS Hold and Wait require that a process request all of its required resources at one time and blocking the process until all requests can be granted simultaneously Mutual Exclusion In general, the first of the four listed conditions cannot be disallowed. If access to a resource requires mutual exclusion, then mutual exclusion must be supported by the OS. Some resources, such as files, may allow multiple accesses for reads but only exclusive access for writes. Even in this case, deadlock can occur if more than one process requires write permission. Hold and Wait The hold-and-wait condition can be prevented by requiring that a process request all of its required resources at one time and blocking the process until all requests can be granted simultaneously. This approach is inefficient in two ways. First, a process may be held up for a long time waiting for all of its resource requests to be filled, when in fact it could have proceeded with only some of the resources. Second, resources allocated to a process may remain unused for a considerable period, during which time they are denied to other processes. Another problem is that a process may not know in advance all of the resources that it will require. There is also the practical problem created by the use of modular programming or a multithreaded structure for an application. An application would need to be aware of all resources that will be requested at all levels or in all modules to make the simultaneous request.

74 Deadlock Condition Prevention No Preemption
if a process holding certain resources is denied a further request, that process must release its original resources and request them again OS may preempt the second process and require it to release its resources Circular Wait define a linear ordering of resource types No Preemption This condition can be prevented in several ways. First, if a process holding certain resources is denied a further request, that process must release its original resources and, if necessary, request them again together with the additional resource. Alternatively, if a process requests a resource that is currently held by another process, the OS may preempt the second process and require it to release its resources. This latter scheme would prevent deadlock only if no two processes possessed the same priority. This approach is practical only when applied to resources whose state can be easily saved and restored later, as is the case with a processor. Circular Wait The circular-wait condition can be prevented by defining a linear ordering of resource types. If a process has been allocated resources of type R , then it may subsequently request only those resources of types following R in the ordering. To see that this strategy works, let us associate an index with each resource type. Then resource R i precedes R j in the ordering if i < j . Now suppose that two processes, A and B, are deadlocked because A has acquired R i and requested R j , and B has acquired R j and requested R i . This condition is impossible because it implies i < j and j < i . As with hold-and-wait prevention, circular-wait prevention may be inefficient, slowing down processes and denying resource access unnecessarily.

75 Deadlock Avoidance A decision is made dynamically whether the current resource allocation request will, if granted, potentially lead to a deadlock Requires knowledge of future process requests An approach to solving the deadlock problem that differs subtly from deadlock prevention is deadlock avoidance. 2 In deadlock prevention , we constrain resource requests to prevent at least one of the four conditions of deadlock. This is either done indirectly, by preventing one of the three necessary policy conditions (mutual exclusion, hold and wait, no preemption), or directly, by preventing circular wait. This leads to inefficient use of resources and inefficient execution of processes. Deadlock avoidance , on the other hand, allows the three necessary conditions but makes judicious choices to assure that the deadlock point is never reached. As such, avoidance allows more concurrency than prevention. With deadlock avoidance, a decision is made dynamically whether the current resource allocation request will, if granted, potentially lead to a deadlock. Deadlock avoidance thus requires knowledge of future process resource requests.

76 Two Approaches to Deadlock Avoidance
Process Initiation Denial do not start a process if its demands might lead to deadlock Resource Allocation Denial do not grant an incremental resource request to a process if this allocation might lead to deadlock In this section, we describe two approaches to deadlock avoidance: • Do not start a process if its demands might lead to deadlock. • Do not grant an incremental resource request to a process if this allocation might lead to deadlock.

77 Resource Allocation Denial
Referred to as the banker’s algorithm State of the system reflects the current allocation of resources to processes Safe state is one in which there is at least one sequence of resource allocations to processes that does not result in a deadlock Unsafe state is a state that is not safe The strategy of resource allocation denial, referred to as the banker’s algorithm , was first proposed in [DIJK65]. Let us begin by defining the concepts of state and safe state. Consider a system with a fixed number of processes and a fixed number of resources. At any time a process may have zero or more resources allocated to it. The state of the system reflects the current allocation of resources to processes. Thus, the state consists of the two vectors, Resource and Available, and the two matrices, Claim and Allocation, defined earlier. A safe state is one in which there is at least one sequence of resource allocations to processes that does not result in a deadlock (i.e., all of the processes can be run to completion). An unsafe state is, of course, a state that is not safe.

78 Determination of a Safe State
State of a system consisting of four processes and three resources Allocations have been made to the four processes Animation: Callouts explain resource parts of figure The following example illustrates these concepts. Figure 6.7a shows the state of a system consisting of four processes and three resources. The total amount of resources R1, R2, and R3 are 9, 3, and 6 units, respectively. In the current state allocations have been made to the four processes, leaving 1 unit of R2 and 1 unit of R3 available. Is this a safe state? To answer this question, we ask an intermediate question: Can any of the four processes be run to completion with the resources available? That is, can the difference between the maximum requirement and current allocation for any process be met with the available resources? In terms of the matrices and vectors introduced earlier, the condition to be met for process i is: Cij - Aij … Vj, for all j Amount of existing resources Resources available after allocation

79 P2 Runs to Completion Clearly, this is not possible for P1, which has only 1 unit of R1 and requires 2 more units of R1, 2 units of R2, and 2 units of R3. However, by assigning one unit of R3 to process P2, P2 has its maximum required resources allocated and can run to completion. Let us assume that this is accomplished. When P2 completes, its resources can be returned to the pool of available resources. The resulting state is shown in Figure 6.7b . Now we can ask again if any of the remaining processes can be completed. In this case, each of the remaining processes could be completed.

80 P1 Runs to Completion Suppose we choose P1, allocate the required resources, complete P1, and return all of P1’s resources to the available pool. We are left in the state shown in Figure 6.7c .

81 Thus, the state defined originally is a safe state
P3 Runs to Completion Next, we can complete P3, resulting in the state of Figure 6.7d . Finally, we can complete P4. At this point, all of the processes have been run to completion. Thus, the state defined by Figure 6.7a is a safe state. These concepts suggest the following deadlock avoidance strategy, which ensures that the system of processes and resources is always in a safe state. When a process makes a request for a set of resources, assume that the request is granted, update the system state accordingly, and then determine if the result is a safe state. If so, grant the request and, if not, block the process until it is safe to grant the request. Thus, the state defined originally is a safe state

82 Determination of an Unsafe State
Consider the state defined in Figure 6.8a . Suppose P2 makes a request for one additional unit of R1 and one additional unit of R3. If we assume the request is granted, then the resulting state is that of Figure 6.7a . We have already seen that this is a safe state; therefore, it is safe to grant the request. Now let us return to the state of Figure 6.8a and suppose that P1 makes the request for one additional unit each of R1 and R3; if we assume that the request is granted, we are left in the state of Figure 6.8b . Is this a safe state? The answer is no, because each process will need at least one additional unit of R1, and there are none available. Thus, on the basis of deadlock avoidance, the request by P1 should be denied and P1 should be blocked. It is important to point out that Figure 6.8b is not a deadlocked state. It merely has the potential for deadlock. It is possible, for example, that if P1 were run from this state it would subsequently release one unit of R1 and one unit of R3 prior to needing these resources again. If that happened, the system would return to a safe state. Thus, the deadlock avoidance strategy does not predict deadlock with certainty; it merely anticipates the possibility of deadlock and assures that there is never such a possibility.

83 Deadlock Avoidance Logic
Figure 6.9 gives an abstract version of the deadlock avoidance logic. The main algorithm is shown in part (b). With the state of the system defined by the data structure state , request[*] is a vector defining the resources requested by process i . First, a check is made to assure that the request does not exceed the original claim of the process. If the request is valid, the next step is to determine if it is possible to fulfill the request (i.e., there are sufficient resources available). If it is not possible, then the process is suspended. If it is possible, the final step is to determine if it is safe to fulfill the request. To do this, the resources are tentatively assigned to process i to form newstate .

84 Deadlock Avoidance Logic
Then a test for safety is made using the algorithm in Figure 6.9c.

85 Deadlock Avoidance Advantages
It is not necessary to preempt and rollback processes, as in deadlock detection It is less restrictive than deadlock prevention Deadlock avoidance has the advantage that it is not necessary to preempt and rollback processes, as in deadlock detection, and is less restrictive than deadlock prevention.

86 Deadlock Avoidance Restrictions
Maximum resource requirement for each process must be stated in advance Processes under consideration must be independent and with no synchronization requirements There must be a fixed number of resources to allocate No process may exit while holding resources However, it does have a number of restrictions on its use: • The maximum resource requirement for each process must be stated in advance. • The processes under consideration must be independent; that is, the order in which they execute must be unconstrained by any synchronization requirements. • There must be a fixed number of resources to allocate. • No process may exit while holding resources.

87 Deadlock prevention strategies are very conservative
Deadlock Strategies Deadlock prevention strategies are very conservative limit access to resources by imposing restrictions on processes Deadlock detection strategies do the opposite resource requests are granted whenever possible Deadlock prevention strategies are very conservative; they solve the problem of deadlock by limiting access to resources and by imposing restrictions on processes. At the opposite extreme, deadlock detection strategies do not limit resource access or restrict process actions. With deadlock detection, requested resources are granted to processes whenever possible. Periodically, the OS performs an algorithm that allows it to detect the circular wait condition described earlier in condition (4) and illustrated in Figure 6.6 .

88 Deadline Detection Algorithms
A check for deadlock can be made as frequently as each resource request or, less frequently, depending on how likely it is for a deadlock to occur Advantages: it leads to early detection the algorithm is relatively simple Disadvantage frequent checks consume considerable processor time A check for deadlock can be made as frequently as each resource request or, less frequently, depending on how likely it is for a deadlock to occur. Checking at each resource request has two advantages: It leads to early detection, and the algorithm is relatively simple because it is based on incremental changes to the state of the system. On the other hand, such frequent checks consume considerable processor time.

89 Deadlock Detection Algorithm
We can use Figure 6.10 to illustrate the deadlock detection algorithm. The algorithm proceeds as follows: 1. Mark P4, because P4 has no allocated resources. 2. Set W = ( ). 3. The request of process P3 is less than or equal to W, so mark P3 and set W=W + ( ) = ( ). 4. No other unmarked process has a row in Q that is less than or equal to W. Therefore, terminate the algorithm. The algorithm concludes with P1 and P2 unmarked, indicating that these processes are deadlocked.

90 Recovery Strategies Abort all deadlocked processes
Back up each deadlocked process to some previously defined checkpoint and restart all processes Successively abort deadlocked processes until deadlock no longer exists Successively preempt resources until deadlock no longer exists Once deadlock has been detected, some strategy is needed for recovery. The following are possible approaches, listed in order of increasing sophistication: 1. Abort all deadlocked processes. This is, believe it or not, one of the most common, if not the most common, solution adopted in operating systems. 2. Back up each deadlocked process to some previously defined checkpoint, and restart all processes. This requires that rollback and restart mechanisms be built in to the system. The risk in this approach is that the original deadlock may recur. However, the nondeterminancy of concurrent processing may ensure that this does not happen. 3. Successively abort deadlocked processes until deadlock no longer exists. The order in which processes are selected for abortion should be on the basis of some criterion of minimum cost. After each abortion, the detection algorithm must be reinvoked to see whether deadlock still exists. 4. Successively preempt resources until deadlock no longer exists. As in (3), a costbased selection should be used, and reinvocation of the detection algorithm is required after each preemption. A process that has a resource preempted from it must be rolled back to a point prior to its acquisition of that resource. For (3) and (4), the selection criteria could be one of the following. Choose the process with the • least amount of processor time consumed so far • least amount of output produced so far • most estimated time remaining least total resources allocated so far • lowest priority Some of these quantities are easier to measure than others. Estimated time remaining is particularly suspect. Also, other than by means of the priority measure, there is no indication of the “cost” to the user, as opposed to the cost to the system as a whole.

91 Deadlock Approaches As Table 6.1 suggests, there are strengths and weaknesses to all of the strategies for dealing with deadlock. Rather than attempting to design an OS facility that employs only one of these strategies, it might be more efficient to use different strategies in different situations.

92 Scheduling

93 medium term scheduling
Processor Scheduling Aim is to assign processes to be executed by the processor in a way that meets system objectives, such as response time, throughput, and processor efficiency Broken down into three separate functions: long term scheduling medium term scheduling short term scheduling The aim of processor scheduling is to assign processes to be executed by the processor or processors over time, in a way that meets system objectives, such as response time, throughput, and processor efficiency. In many systems, this scheduling activity is broken down into three separate functions: long-, medium-, and short-term scheduling. The names suggest the relative time scales with which these functions are performed.

94 Table 9.1 Types of Scheduling
The key to multiprogramming is scheduling. In fact, four types of scheduling are typically involved ( Table 9.1 ). One of these, I/O scheduling, is more conveniently addressed in Chapter 11 , where I/O is discussed. The remaining three types of scheduling, which are types of processor scheduling, are addressed in this chapter and the next.

95 Scheduling and Process State Transitions
Figure 9.1 relates the scheduling functions to the process state transition diagram (first shown in Figure 3.9b ). Long-term scheduling is performed when a new process is created. This is a decision whether to add a new process to the set of processes that are currently active. Medium-term scheduling is a part of the swapping function. This is a decision whether to add a process to those that are at least partially in main memory and therefore available for execution. Short-term scheduling is the actual decision of which ready process to execute next.

96 Figure 9.2 Nesting of Scheduling Functions
(Referencing figure 3.9b) Figure 9.2 reorganizes the state transition diagram of Figure 3.9b to suggest the nesting of scheduling functions.

97 Queuing Diagram Scheduling affects the performance of the system because it determines which processes will wait and which will progress. This point of view is presented in Figure 9.3 , which shows the queues involved in the state transitions of a process. 1 Fundamentally, scheduling is a matter of managing queues to minimize queuing delay and to optimize performance in a queuing environment.

98 Long-Term Scheduler Determines which programs are admitted to the system for processing Controls the degree of multiprogramming the more processes that are created, the smaller the percentage of time that each process can be executed may limit to provide satisfactory service to the current set of processes Creates processes from the queue when it can, but must decide: when the operating system can take on one or more additional processes which jobs to accept and turn into processes first come, first served priority, expected execution time, I/O requirements The long-term scheduler determines which programs are admitted to the system for processing. Thus, it controls the degree of multiprogramming. Once admitted, a job or user program becomes a process and is added to the queue for the short-term scheduler. In some systems, a newly created process begins in a swapped-out condition, in which case it is added to a queue for the medium-term scheduler. In a batch system, or for the batch portion of a general-purpose operating system, newly submitted jobs are routed to disk and held in a batch queue. The long-term scheduler creates processes from the queue when it can. There are two decisions involved here. First, the scheduler must decide when the operating system can take on one or more additional processes. Second, the scheduler must decide which job or jobs to accept and turn into processes. The decision as to when to create a new process is generally driven by the desired degree of multiprogramming. The more processes that are created, the smaller is the percentage of time that each process can be executed (i.e., more processes are competing for the same amount of processor time). Thus, the long-term scheduler may limit the degree of multiprogramming to provide satisfactory service to the current set of processes. Each time a job terminates, the scheduler may decide to add one or more new jobs. Additionally, if the fraction of time that the processor is idle exceeds a certain threshold, the long-term scheduler may be invoked. The decision as to which job to admit next can be on a simple first-comefirst- served (FCFS) basis, or it can be a tool to manage system performance. The criteria used may include priority, expected execution time, and I/O requirements. For example, if the information is available, the scheduler may attempt to keep a mix of processor-bound and I/O-bound processes. 2 Also, the decision can depend on which I/O resources are to be requested, in an attempt to balance I/O usage. For interactive programs in a time-sharing system, a process creation request can be generated by the act of a user attempting to connect to the system. Timesharing users are not simply queued up and kept waiting until the system can accept them. Rather, the operating system will accept all authorized comers until the system is saturated, using some predefined measure of saturation. At that point, a connection request is met with a message indicating that the system is full and the user should try again later.

99 Medium-Term Scheduling
Part of the swapping function Swapping-in decisions are based on the need to manage the degree of multiprogramming considers the memory requirements of the swapped-out processes Medium-term scheduling is part of the swapping function. The issues involved are discussed in Chapters 3 , 7 , and 8 . Typically, the swapping-in decision is based on the need to manage the degree of multiprogramming. On a system that does not use virtual memory, memory management is also an issue. Thus, the swapping-in decision will consider the memory requirements of the swapped-out processes.

100 Short-Term Scheduling
Known as the dispatcher Executes most frequently Makes the fine-grained decision of which process to execute next Invoked when an event occurs that leads to the blocking of the current process or that may provide an opportunity to preempt a currently running process in favor of another In terms of frequency of execution, the long-term scheduler executes relatively infrequently and makes the coarse-grained decision of whether or not to take on a new process and which one to take. The medium-term scheduler is executed somewhat more frequently to make a swapping decision. The short-term scheduler, also known as the dispatcher, executes most frequently and makes the fine-grained decision of which process to execute next. The short-term scheduler is invoked whenever an event occurs that may lead to the blocking of the current process or that may provide an opportunity to preempt a currently running process in favor of another. Examples of such events include • Clock interrupts • I/O interrupts • Operating system calls • Signals (e.g., semaphores) Examples: Clock interrupts I/O interrupts Operating system calls Signals (e.g., semaphores)

101 Short Term Scheduling Criteria
Main objective is to allocate processor time to optimize certain aspects of system behavior A set of criteria is needed to evaluate the scheduling policy User-oriented criteria relate to the behavior of the system as perceived by the individual user or process (such as response time in an interactive system) important on virtually all systems System-oriented criteria focus in on effective and efficient utilization of the processor (rate at which processes are completed) generally of minor importance on single-user systems The main objective of short-term scheduling is to allocate processor time in such a way as to optimize one or more aspects of system behavior. Generally, a set of criteria is established against which various scheduling policies may be evaluated. The commonly used criteria can be categorized along two dimensions. First, we can make a distinction between user-oriented and system-oriented criteria. User-oriented criteria relate to the behavior of the system as perceived by the individual user or process. An example is response time in an interactive system. Response time is the elapsed time between the submission of a request until the response begins to appear as output. This quantity is visible to the user and is naturally of interest to the user. We would like a scheduling policy that provides “good” service to various users. In the case of response time, a threshold may be defined, say two seconds. Then a goal of the scheduling mechanism should be to maximize the number of users who experience an average response time of two seconds or less. Other criteria are system oriented. That is, the focus is on effective and efficient utilization of the processor. An example is throughput, which is the rate at which processes are completed. This is certainly a worthwhile measure of system performance and one that we would like to maximize. However, it focuses on system performance rather than service provided to the user. Thus, throughput is of concern to a system administrator but not to the user population. Whereas user-oriented criteria are important on virtually all systems, systemoriented criteria are generally of minor importance on single-user systems. On a single-user system, it probably is not important to achieve high processor utilization or high throughput as long as the responsiveness of the system to user applications is acceptable.

102 Short-Term Scheduling Criteria: Performance
Criteria can be classified into: Performance-related quantitative easily measured Non-performance related qualitative hard to measure example: predictability examples: response time throughput Another dimension along which criteria can be classified is those that are performance related and those that are not directly performance related. Performance-related criteria are quantitative and generally can be readily measured. Examples include response time and throughput. Criteria that are not performance related are either qualitative in nature or do not lend themselves readily to measurement and analysis. An example of such a criterion is predictability. We would like for the service provided to users to exhibit the same characteristics over time, independent of other work being performed by the system. To some extent, this criterion can be measured by calculating variances as a function of workload. However, this is not nearly as straightforward as measuring throughput or response time as a function of workload.

103 Table 9.2 Scheduling Criteria
Table 9.2 summarizes key scheduling criteria. These are interdependent, and it is impossible to optimize all of them simultaneously. For example, providing good response time may require a scheduling algorithm that switches between processes frequently. This increases the overhead of the system, reducing throughput. Thus, the design of a scheduling policy involves compromising among competing requirements; the relative weights given the various requirements will depend on the nature and intended use of the system. In most interactive operating systems, whether single user or time shared, adequate response time is the critical requirement. Because of the importance of this requirement, and because the definition of adequacy will vary from one application to another, the topic is explored further in Appendix G .

104 Priority Queuing In many systems, each process is assigned a priority and the scheduler will always choose a process of higher priority over one of lower priority. Figure 9.4 illustrates the use of priorities. For clarity, the queuing diagram is simplified, ignoring the existence of multiple blocked queues and of suspended states (compare Figure 3.8a ). Instead of a single ready queue, we provide a set of queues, in descending order of priority: RQ0, RQ1, , RQ n , with priority[RQ i ] > priority[RQ j ] for i > j . 3 When a scheduling selection is to be made, the scheduler will start at the highest-priority ready queue (RQ0). If there are one or more processes in the queue, a process is selected using some scheduling policy. If RQ0 is empty, then RQ1 is examined, and so on. One problem with a pure priority scheduling scheme is that lower-priority processes may suffer starvation. This will happen if there is always a steady supply of higher-priority ready processes. If this behavior is not desirable, the priority of a process can change with its age or execution history. We will give one example of this subsequently.

105 Selection Function Determines which Ready process is dispatched next
May be based on priority, resource requirements, or the execution characteristics of the process If based on execution characteristics, some factors to consider are w = time spent in system so far, waiting e = time spent in execution so far s = total service time required by the process, including e; (estimated by system or user) . The selection function determines which process, among ready processes, is selected next for execution. The function may be based on priority, resource requirements, or the execution characteristics of the process. In the latter case, three quantities are significant: w = time spent in system so far, waiting e = time spent in execution so far s = total service time required by the process, including e ; generally, this quantity must be estimated or supplied by the user For example, the selection function max[ w ] indicates an FCFS discipline.

106 When/under what circumstances is the selection function is exercised?
Decision Mode Two categories: Nonpreemptive Preemptive When/under what circumstances is the selection function is exercised? The decision mode specifies the instants in time at which the selection function is exercised. There are two general categories: nonpreemptive and preemptive

107 Nonpreemptive vs Preemptive
once a process is in the running state, it will continue until it terminates or blocks itself for I/O currently running process may be interrupted and moved to ready state by the OS preemption may occur when a new process arrives, on an interrupt, or periodically Non-preemptive In this case, once a process is in the Running state, it continues to execute until it terminates or it blocks itself to wait for I/O or to request some operating system service. Preemptive: The currently running process may be interrupted and moved to the Ready state by the operating system. The decision to preempt may be performed when a new process arrives; when an interrupt occurs that places a blocked process in the Ready state; or periodically, based on a clock interrupt.

108 Alternative Scheduling Policies
Table 9.3 presents some summary information about the various scheduling policies that are examined in this subsection

109 Table 9.4 Process Scheduling Example
As we describe the various scheduling policies, we will use the set of processes in Table 9.4 as a running example. We can think of these as batch jobs, with the service time being the total execution time required. Alternatively, we can consider these to be ongoing processes that require alternate use of the processor and I/O in a repetitive fashion. In this latter case, the service times represent the processor time required in one cycle. In either case, in terms of a queuing model, this quantity corresponds to the service time.

110 Table 9.5 Comparison of Scheduling Policies
For the example of Table 9.4 , Figure 9.5 shows the execution pattern for each policy for one cycle, and Table 9.5 summarizes some key results. First, the finish time of each process is determined. From this, we can determine the turnaround time. In terms of the queuing model, turnaround time (TAT) is the residence time T r , or total time that the item spends in the system (waiting time plus service time). A more useful figure is the normalized turnaround time, which is the ratio of turnaround time to service time. This value indicates the relative delay experienced by a process. Typically, the longer the process execution time, the greater is the absolute amount of delay that can be tolerated. The minimum possible value for this ratio is 1.0; increasing values correspond to a decreasing level of service. (Assumes no process blocks itself, for I/O or other event wait.)

111 First-Come-First-Served (FCFS)
Simplest scheduling policy Also known as first-in-first-out (FIFO) or a strict queuing scheme When the current process ceases to execute, the longest process in the Ready queue is selected Performs much better for long processes than short ones Tends to favor processor-bound processes over I/O-bound processes The simplest scheduling policy is first-come-first-served (FCFS), also known as first-in-first-out (FIFO) or a strict queuing scheme. As each process becomes ready, it joins the ready queue. When the currently running process ceases to execute, the process that has been in the ready queue the longest is selected for running. FCFS performs much better for long processes than short ones. Another difficulty with FCFS is that it tends to favor processor-bound processes over I/O-bound processes. Consider that there is a collection of processes, one of which mostly uses the processor (processor bound) and a number of which favor I/O (I/O bound). When a processor-bound process is running, all of the I/O bound processes must wait. Some of these may be in I/O queues (blocked state) but may move back to the ready queue while the processor-bound process is executing. At this point, most or all of the I/O devices may be idle, even though there is potentially work for them to do. When the currently running process leaves the Running state, the ready I/O-bound processes quickly move through the Running state and become blocked on I/O events. If the processor-bound process is also blocked, the processor becomes idle. Thus, FCFS may result in inefficient use of both the processor and the I/O devices. FCFS is not an attractive alternative on its own for a uniprocessor system. However, it is often combined with a priority scheme to provide an effective scheduler. Thus, the scheduler may maintain a number of queues, one for each priority level, and dispatch within each queue on a first-come-first-served basis. We see one example of such a system later, in our discussion of feedback scheduling.

112 Round Robin Uses preemption based on a clock
Also known as time slicing because each process is given a slice of time before being preempted Principal design issue is the length of the time quantum, or slice, to be used Particularly effective in a general-purpose time-sharing system or transaction processing system One drawback is its relative treatment of processor-bound and I/O-bound processes A straightforward way to reduce the penalty that short jobs suffer with FCFS is to use preemption based on a clock. The simplest such policy is round robin. A clock interrupt is generated at periodic intervals. When the interrupt occurs, the currently running process is placed in the ready queue, and the next ready job is selected on a FCFS basis. This technique is also known as time slicing , because each process is given a slice of time before being preempted. With round robin, the principal design issue is the length of the time quantum, or slice, to be used. If the quantum is very short, then short processes will move through the system relatively quickly. On the other hand, there is processing overhead involved in handling the clock interrupt and performing the scheduling and dispatching function. Thus, very short time quanta should be avoided. Round robin is particularly effective in a general-purpose time-sharing system or transaction processing system. One drawback to round robin is its relative treatment of processor-bound and I/O-bound processes. Generally, an I/O-bound process has a shorter processor burst (amount of time spent executing between I/O operations) than a processor-bound process. If there is a mix of processor-bound and I/O-bound processes, then the following will happen: An I/O-bound process uses a processor for a short period and then is blocked for I/O; it waits for the I/O operation to complete and then joins the ready queue. On the other hand, a processor-bound process generally uses a complete time quantum while executing and immediately returns to the ready queue. Thus, processor-bound processes tend to receive an unfair portion of processor time, which results in poor performance for I/O-bound processes, inefficient use of I/O devices, and an increase in the variance of response time.

113 Effect of Size of Preemption Time Quantum
Figure 9.6a Effect of Size of Preemption Time Quantum One useful guide is that the time quantum should be slightly greater than the time required for a typical interaction or process function. If it is less, then most processes will require at least two time quanta. Figure 9.6 illustrates the effect this has on response time. Note that in the limiting case of a time quantum that is longer than the longest running process, round robin degenerates to FCFS.

114 Figure 9.6b Effect of Size of Preemption Time Quantum
Figure 9.6 continued. Effect of size of preemption time quantum.

115 Virtual Round Robin (VRR)
[HALD91] suggests a refinement to round robin that he refers to as a virtual round robin (VRR) and that avoids this unfairness. Figure 9.7 illustrates the scheme. New processes arrive and join the ready queue, which is managed on an FCFS basis. When a running process times out, it is returned to the ready queue. When a process is blocked for I/O, it joins an I/O queue. So far, this is as usual. The new feature is an FCFS auxiliary queue to which processes are moved after being released from an I/O block. When a dispatching decision is to be made, processes in the auxiliary queue get preference over those in the main ready queue. When a process is dispatched from the auxiliary queue, it runs no longer than a time equal to the basic time quantum minus the total time spent running since it was last selected from the main ready queue. Performance studies by the authors indicate that this approach is indeed superior to round robin in terms of fairness.

116 Shortest Process Next (SPN)
Nonpreemptive policy in which the process with the shortest expected processing time is selected next A short process will jump to the head of the queue Possibility of starvation for longer processes One difficulty is the need to know, or at least estimate, the required processing time of each process If the programmer’s estimate is substantially under the actual running time, the system may abort the job Another approach to reducing the bias in favor of long processes inherent in FCFS is the shortest process next (SPN) policy. This is a nonpreemptive policy in which the process with the shortest expected processing time is selected next. Thus, a short process will jump to the head of the queue past longer jobs. Figure 9.5 and Table 9.5 show the results for our example. Note that process E receives service much earlier than under FCFS. Overall performance is also significantly improved in terms of response time. However, the variability of response times is increased, especially for longer processes, and thus predictability is reduced. One difficulty with the SPN policy is the need to know or at least estimate the required processing time of each process. For batch jobs, the system may require the programmer to estimate the value and supply it to the operating system. If the programmer’s estimate is substantially under the actual running time, the system may abort the job. In a production environment, the same jobs run frequently, and statistics may be gathered. For interactive processes, the operating system may keep a running average of each “burst” for each process. A risk with SPN is the possibility of starvation for longer processes, as long as there is a steady supply of shorter processes. On the other hand, although SPN reduces the bias in favor of longer jobs, it still is not desirable for a time-sharing or transaction processing environment because of the lack of preemption. Looking back at our worst-case analysis described under FCFS, processes W, X, Y, and Z will still execute in the same order, heavily penalizing the short process Y.

117 Shortest Process Next (SPN)
Problem: Estimating execution time OS may collect statistics and use process history to estimate run time e.g., for processes in a production environment Problem: avoiding starvation for long processes Problem: not suitable for timesharing or transaction processing due to no preemption. The shortest remaining time (SRT) policy is a preemptive version of SPN. In this case, the scheduler always chooses the process that has the shortest expected remaining processing time. When a new process joins the ready queue, it may in fact have a shorter remaining time than the currently running process. Accordingly, the scheduler may preempt the current process when a new process becomes ready. As with SPN, the scheduler must have an estimate of processing time to perform the selection function, and there is a risk of starvation of longer processes. SRT does not have the bias in favor of long processes found in FCFS. Unlike round robin, no additional interrupts are generated, reducing overhead. On the other hand, elapsed service times must be recorded, contributing to overhead. SRT should also give superior turnaround time performance to SPN, because a short job is given immediate preference to a running longer job. Note that in our example ( Table 9.5 ), the three shortest processes all receive immediate service, yielding a normalized turnaround time for each of 1.0.

118 Shortest Remaining Time (SRT)
Preemptive version of SPN Scheduler always chooses the process that has the shortest expected remaining processing time Risk of starvation of longer processes Should give superior turnaround time performance to SPN because a short job is given immediate preference to a running longer job Still depends on having accurate service time estimates. The shortest remaining time (SRT) policy is a preemptive version of SPN. In this case, the scheduler always chooses the process that has the shortest expected remaining processing time. When a new process joins the ready queue, it may in fact have a shorter remaining time than the currently running process. Accordingly, the scheduler may preempt the current process when a new process becomes ready. As with SPN, the scheduler must have an estimate of processing time to perform the selection function, and there is a risk of starvation of longer processes. SRT does not have the bias in favor of long processes found in FCFS. Unlike round robin, no additional interrupts are generated, reducing overhead. On the other hand, elapsed service times must be recorded, contributing to overhead. SRT should also give superior turnaround time performance to SPN, because a short job is given immediate preference to a running longer job. Note that in our example ( Table 9.5 ), the three shortest processes all receive immediate service, yielding a normalized turnaround time for each of 1.0.

119 Highest Response Ratio Next (HRRN)
Chooses next process with the greatest ratio Attractive because it accounts for the age of the process While shorter jobs are favored, aging without service increases the ratio so that a longer process will eventually get past competing shorter jobs In Table 9.5 , we have used the normalized turnaround time, which is the ratio of turnaround time to actual service time, as a figure of merit. For each individual process, we would like to minimize this ratio, and we would like to minimize the average value over all processes. In general, we cannot know ahead of time what the service time is going to be, but we can approximate it, either based on past history or some input from the user or a configuration manager. Thus, our scheduling rule becomes the following: when the current process completes or is blocked, choose the ready process with the greatest value of R . This approach is attractive because it accounts for the age of the process. While shorter jobs are favored (a smaller denominator yields a larger ratio), aging without service increases the ratio so that a longer process will eventually get past competing shorter jobs. As with SRT and SPN, the expected service time must be estimated to use highest response ratio next (HRRN).

120 Multilevel Feedback Scheduling
Scheduling is similar to RR: FCFS with a time quantum. However, when a process blocks or is preempted it is “fed back” into the next lower level queue. Once it reaches the lowest level queue a process is served by RR until it terminates. Process is dispatched from the highest priority non-empty queue Result: new processes favored over long older processes Modifications address starvation and I/O bound processes Multilevel Feedback Scheduling Useful when there is no information about relative length of various jobs, but You would like to favor short jobs Basic algorithm may starve long processes or I/O bound processes.

121 Feedback Scheduling Figure 9.10 illustrates the feedback scheduling mechanism by showing the path that a process will follow through the various queues. 5 This approach is known as multilevel feedback , meaning that the operating system allocates the processor to a process and, when the process blocks or is preempted, feeds it back into one of several priority queues.

122 Feedback Performance There are a number of variations on this scheme. A simple version is to perform preemption in the same fashion as for round robin: at periodic intervals. Our example shows this ( Figure 9.5 and Table 9.5 ) for a quantum of one time unit. Note that in this case, the behavior is similar to round robin with a time quantum of 1. One problem with the simple scheme just outlined is that the turnaround time of longer processes can stretch out alarmingly. Indeed, it is possible for starvation to occur if new jobs are entering the system frequently. To compensate for this, we can vary the preemption times according to the queue: A process scheduled from RQ0 is allowed to execute for one time unit and then is preempted; a process scheduled from RQ1 is allowed to execute two time units, and so on. In general, a process scheduled from RQ i is allowed to execute 2 i time units before preemption. This scheme is illustrated for our example in Figure 9.5 and Table 9.5 . Even with the allowance for greater time allocation at lower priority, a longer process may still suffer starvation. A possible remedy is to promote a process to a higher-priority queue after it spends a certain amount of time waiting for service in its current queue.

123 Feedback Queue Modifications
To avoid unreasonably long waits for long processes, give processes in lower-priority queues longer quantums To avoid starvation let a process that has not executed for a certain amount of time move to a higher level queue. To lessen the penalty on I/O bound processes use some version of virtual RR for processes that don’t receive a full quantum.

124 Performance Comparison
Any scheduling discipline that chooses the next item to be served independent of service time obeys the relationship: In this section, we make use of basic queuing formulas, with the common assumptions of Poisson arrivals and exponential service times. 6 First, we make the observation that any such scheduling discipline that chooses the next item to be served independent of service time obeys the following relationship: (see formula) In particular, a priority-based scheduler, in which the priority of each process is assigned independent of expected service time, provides the same average turnaround time and average normalized turnaround time as a simple FCFS discipline. Furthermore, the presence or absence of preemption makes no differences in these averages. With the exception of round robin and FCFS, the various scheduling disciplines considered so far do make selections on the basis of expected service time. Unfortunately, it turns out to be quite difficult to develop closed analytic models of these disciplines. However, we can get an idea of the relative performance of such scheduling algorithms, compared to FCFS, by considering priority scheduling in which priority is based on service time.

125 Formulas for Single-Server Queues with Two Priority Categories
Table 9.6 Formulas for Single-Server Queues with Two Priority Categories If scheduling is done on the basis of priority and if processes are assigned to a priority class on the basis of service time, then differences do emerge. Table 9.6 shows the formulas that result when we assume two priority classes, with different service times for each class. In the table, refers to the arrival rate. These results can be generalized to any number of priority classes. Note that the formulas differ for nonpreemptive versus preemptive scheduling. In the latter case, it is assumed that a lower-priority process is immediately interrupted when a higher-priority process becomes ready.

126 Overall Normalized Response Time
As an example, let us consider the case of two priority classes, with an equal number of process arrivals in each class and with the average service time for the lower-priority class being five times that of the upper priority class. Thus, we wish to give preference to shorter processes. Figure 9.11 shows the overall result. By giving preference to shorter jobs, the average normalized turnaround time is improved at higher levels of utilization. As might be expected, the improvement is greatest with the use of preemption. Notice, however, that overall performance is not much affected.

127 Normalized Response Time for Shorter Processes
However, significant differences emerge when we consider the two priority classes separately. Figure 9.12 shows the results for the higher-priority, shorter processes. For comparison, the upper line on the graph assumes that priorities are not used but that we are simply looking at the relative performance of that half of all processes that have the shorter processing time. The other two lines assume that these processes are assigned a higher priority. When the system is run using priority scheduling without preemption, the improvements are significant. They are even more significant when preemption is used.

128 Normalized Response Time for Longer Processes
Figure 9.13 shows the same analysis for the lower-priority, longer processes. As expected, such processes suffer a performance degradation under priority scheduling.

129 Simulation Results Some of the difficulties of analytic modeling are
overcome by using discrete-event simulation, which allows a wide range of policies to be modeled. The disadvantage of simulation is that the results for a given “run” only apply to that particular collection of processes under that particular set of assumptions. Nevertheless, useful insights can be gained. The results of one such study are reported in [FINK88]. The simulation involved 50,000 processes with an arrival rate of and an average service time of T s 1. Thus, the assumption is that the processor utilization is r = T s = 0.8. Note, therefore, that we are only measuring one utilization point. To present the results, processes are grouped into service-time percentiles, each of which has 500 processes. Thus, the 500 processes with the shortest service time are in the first percentile; with these eliminated, the 500 remaining processes with the shortest service time are in the second percentile; and so on. This allows us to view the effect of various policies on processes as a function of the length of the process. Figure 9.14 shows the normalized turnaround time, and Figure 9.15 shows the average waiting time. Looking at the turnaround time, we can see that the performance of FCFS is very unfavorable, with one-third of the processes having a normalized turnaround time greater than 10 times the service time; furthermore, these are the shortest processes. On the other hand, the absolute waiting time is uniform, as is to be expected because scheduling is independent of service time. The figures show round robin using a quantum of one time unit. Except for the shortest processes, which execute in less than one quantum, round robin yields a normalized turnaround time of about five for all processes, treating all fairly. Shortest process next performs better than round robin, except for the shortest processes. Shortest remaining time, the preemptive version of SPN, performs better than SPN except for the longest 7% of all processes. We have seen that, among nonpreemptive policies, FCFS favors long processes and SPN favors short ones. Highest response ratio next is intended to be a compromise between these two effects, and this is indeed confirmed in the figures. Finally, the figure shows feedback scheduling with fixed, uniform quanta in each priority queue. As expected, FB performs quite well for short processes. Simulation

130 Fair-Share Scheduling
Scheduling decisions based on the process sets Each user is assigned a share of the processor Objective is to monitor usage to give fewer resources to users who have had more than their fair share and more to those who have had less than their fair share All of the scheduling algorithms discussed so far treat the collection of ready processes as a single pool of processes from which to select the next running process. This pool may be broken down by priority but is otherwise homogeneous. However, in a multiuser system, if individual user applications or jobs may be organized as multiple processes (or threads), then there is a structure to the collection of processes that is not recognized by a traditional scheduler. From the user’s point of view, the concern is not how a particular process performs but rather how his or her set of processes, which constitute a single application, performs. Thus, it would be attractive to make scheduling decisions on the basis of these process sets. This approach is generally known as fair-share scheduling. Further, the concept can be extended to groups of users, even if each user is represented by a single process. For example, in a time-sharing system, we might wish to consider all of the users from a given department to be members of the same group. Scheduling decisions could then be made that attempt to give each group similar service. Thus, if a large number of people from one department log onto the system, we would like to see response time degradation primarily affect members of that department rather than users from other departments. The term fair share indicates the philosophy behind such a scheduler. Each user is assigned a weighting of some sort that defines that user’s share of system resources as a fraction of the total usage of those resources. In particular, each user is assigned a share of the processor. Such a scheme should operate in a more or less linear fashion, so that if user A has twice the weighting of user B, then in the long run, user A should be able to do twice as much work as user B. The objective of a fair-share scheduler is to monitor usage to give fewer resources to users who have had more than their fair share and more to those who have had less than their fair share.

131 Fair-Share Scheduler Figure 9.16 is an example in which process A is in one group and processes B and C are in a second group, with each group having a weighting of 0.5. Assume that all processes are processor bound and are usually ready to run. All processes have a base priority of 60. Processor utilization is measured as follows: The processor is interrupted 60 times per second; during each interrupt, the processor usage field of the currently running process is incremented, as is the corresponding group processor field. Once per second, priorities are recalculated. In the figure, process A is scheduled first. At the end of one second, it is preempted. Processes B and C now have the higher priority, and process B is scheduled. At the end of the second time unit, process A has the highest priority. Note that the pattern repeats: the kernel schedules the processes in order: A, B, A, C, A, B, and so on. Thus, 50% of the processor is allocated to process A, which constitutes one group, and 50% to processes B and C, which constitute another group.

132 Traditional UNIX Scheduling
Used in both SVR3 and 4.3 BSD UNIX these systems were primarily targeted at the time-sharing interactive environment Designed to provide good response time for interactive users while ensuring that low-priority background jobs do not starve Employed multilevel feedback using round robin within each of the priority queues Made use of one-second preemption Priority based on process type and execution history In this section we examine traditional UNIX scheduling, which is used in both SVR3 and 4.3 BSD UNIX. These systems are primarily targeted at the time-sharing interactive environment. The scheduling algorithm is designed to provide good response time for interactive users while ensuring that low-priority background jobs do not starve. Although this algorithm has been replaced in modern UNIX systems, it is worthwhile to examine the approach because it is representative of practical time-sharing scheduling algorithms. The scheduling scheme for SVR4 includes an accommodation for real-time requirements, and so its discussion is deferred to Chapter 10 . The traditional UNIX scheduler employs multilevel feedback using round robin within each of the priority queues. The system makes use of one-second preemption. That is, if a running process does not block or complete within one second, it is preempted. Priority is based on process type and execution history.

133 Scheduling Formula The priority of each process is recomputed once per second, at which time a new scheduling decision is made. The purpose of the base priority is to divide all processes into fixed bands of priority levels. The CPU and nice components are restricted to prevent a process from migrating out of its assigned band (assigned by the base priority level). These bands are used to optimize access to block devices (e.g., disk) and to allow the operating system to respond quickly to system calls.

134 Bands Swapper Block I/O device control File manipulation Character I/O device control User processes Used to optimize access to block devices and to allow the operating system to respond quickly to system calls In decreasing order of priority, the bands are: In decreasing order of priority, the bands are: • Swapper • Block I/O device control • File manipulation • Character I/O device control • User processes This hierarchy should provide the most efficient use of the I/O devices. Within the user process band, the use of execution history tends to penalize processor- bound processes at the expense of I/O-bound processes. Again, this should improve efficiency. Coupled with the round-robin preemption scheme, the scheduling strategy is well equipped to satisfy the requirements for general-purpose time sharing.

135 Example of Traditional UNIX Process Scheduling
An example of process scheduling is shown in Figure Processes A, B, and C are created at the same time with base priorities of 60 (we will ignore the nice value). The clock interrupts the system 60 times per second and increments a counter for the running process. The example assumes that none of the processes block themselves and that no other processes are ready to run. Compare this with Figure


Download ppt "Concurrency: Mutual Exclusion and Synchronization"

Similar presentations


Ads by Google