A Sophomoric Introduction to Shared-Memory Parallelism and Concurrency Lecture 4 Shared-Memory Concurrency & Mutual Exclusion Dan Grossman Last Updated:

Slides:



Advertisements
Similar presentations
Implementation and Verification of a Cache Coherence protocol using Spin Steven Farago.
Advertisements

Chapter 6: Process Synchronization
Mutual Exclusion.
Chapter 3 The Critical Section Problem
Parallel Processing (CS526) Spring 2012(Week 6).  A parallel algorithm is a group of partitioned tasks that work with each other to solve a large problem.
Section 6: Mutual Exclusion Michelle Kuttel
Selections from CSE332 Data Abstractions course at University of Washington 1.Introduction to Multithreading & Fork-Join Parallelism –hashtable, vector.
Operating Systems ECE344 Ding Yuan Synchronization (I) -- Critical region and lock Lecture 5: Synchronization (I) -- Critical region and lock.
1 Friday, June 16, 2006 "In order to maintain secrecy, this posting will self-destruct in five seconds. Memorize it, then eat your computer." - Anonymous.
CS444/CS544 Operating Systems Synchronization 2/16/2006 Prof. Searleman
CSE332: Data Abstractions Lecture 22: Shared-Memory Concurrency and Mutual Exclusion Tyler Robison Summer
CSE332: Data Abstractions Lecture 22: Shared-Memory Concurrency and Mutual Exclusion Dan Grossman Spring 2010.
Threading Part 2 CS221 – 4/22/09. Where We Left Off Simple Threads Program: – Start a worker thread from the Main thread – Worker thread prints messages.
A Sophomoric Introduction to Shared-Memory Parallelism and Concurrency Lecture 4 Shared-Memory Concurrency & Mutual Exclusion Dan Grossman Last Updated:
1 Sharing Objects – Ch. 3 Visibility What is the source of the issue? Volatile Dekker’s algorithm Publication and Escape Thread Confinement Immutability.
A. Frank - P. Weisberg Operating Systems Introduction to Cooperating Processes.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Emery Berger University of Massachusetts, Amherst Operating Systems CMPSCI 377 Lecture.
Synchronization CSCI 444/544 Operating Systems Fall 2008.
02/19/2007CSCI 315 Operating Systems Design1 Process Synchronization Notice: The slides for this lecture have been largely based on those accompanying.
Process Synchronization Ch. 4.4 – Cooperating Processes Ch. 7 – Concurrency.
Concurrency Recitation – 2/24 Nisarg Raval Slides by Prof. Landon Cox.
1 CSCD 330 Network Programming Lecture 13 More Client-Server Programming Sometime in 2014 Reading: References at end of Lecture.
1 Thread II Slides courtesy of Dr. Nilanjan Banerjee.
A Sophomoric Introduction to Shared-Memory Parallelism and Concurrency Lecture 5 Programming with Locks and Critical Sections Original Work by: Dan Grossman.
Games Development 2 Concurrent Programming CO3301 Week 9.
Optimistic Design 1. Guarded Methods Do something based on the fact that one or more objects have particular states  Make a set of purchases assuming.
COMP 111 Threads and concurrency Sept 28, Tufts University Computer Science2 Who is this guy? I am not Prof. Couch Obvious? Sam Guyer New assistant.
Operating Systems ECE344 Ashvin Goel ECE University of Toronto Mutual Exclusion.
Internet Software Development Controlling Threads Paul J Krause.
Producer-Consumer Problem The problem describes two processes, the producer and the consumer, who share a common, fixed-size buffer used as a queue.bufferqueue.
CSC321 Concurrent Programming: §5 Monitors 1 Section 5 Monitors.
11/18/20151 Operating Systems Design (CS 423) Elsa L Gunter 2112 SC, UIUC Based on slides by Roy Campbell, Sam.
Lecture 20: Parallelism & Concurrency CS 62 Spring 2013 Kim Bruce & Kevin Coogan CS 62 Spring 2013 Kim Bruce & Kevin Coogan Some slides based on those.
U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science Software Systems Advanced Synchronization Emery Berger and Mark Corner University.
CS399 New Beginnings Jonathan Walpole. 2 Concurrent Programming & Synchronization Primitives.
Monitors and Blocking Synchronization Dalia Cohn Alperovich Based on “The Art of Multiprocessor Programming” by Herlihy & Shavit, chapter 8.
U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science Computer Systems Principles Synchronization Emery Berger and Mark Corner University.
Fall 2008Programming Development Techniques 1 Topic 20 Concurrency Section 3.4.
Lecture 3 Concurrency and Thread Synchronization     Mutual Exclusion         Dekker's Algorithm         Lamport's Bakery Algorithm.
1 Previous Lecture Overview  semaphores provide the first high-level synchronization abstraction that is possible to implement efficiently in OS. This.
CS 153 Design of Operating Systems Winter 2016 Lecture 7: Synchronization.
CPE779: More on OpenMP Based on slides by Laxmikant V. Kale and David Padua of the University of Illinois.
Concurrency (Threads) Threads allow you to do tasks in parallel. In an unthreaded program, you code is executed procedurally from start to finish. In a.
Implementing Mutual Exclusion Andy Wang Operating Systems COP 4610 / CGS 5765.
Synchronization Questions answered in this lecture: Why is synchronization necessary? What are race conditions, critical sections, and atomic operations?
Mergesort example: Merge as we return from recursive calls Merge Divide 1 element 829.
6/27/20161 Operating Systems Design (CS 423) Elsa L Gunter 2112 SC, UIUC Based on slides by Roy Campbell, Sam King,
CSCD 330 Network Programming
Multithreading / Concurrency
CSE 120 Principles of Operating
Background on the need for Synchronization
Shared-Memory Concurrency & Mutual Exclusion
Atomic Operations in Hardware
Atomic Operations in Hardware
CSE 332: Locks and Deadlocks
CSE332: Data Abstractions Lecture 21: Shared-Memory Concurrency & Mutual Exclusion Dan Grossman Spring 2012.
Liveness And Performance
Multithreading.
Implementing Mutual Exclusion
Implementing Mutual Exclusion
CSE 451: Operating Systems Autumn 2004 Module 6 Synchronization
CSE 451: Operating Systems Autumn 2003 Lecture 7 Synchronization
CSE 451: Operating Systems Autumn 2005 Lecture 7 Synchronization
CSE 451: Operating Systems Winter 2003 Lecture 7 Synchronization
CSE 153 Design of Operating Systems Winter 19
CS333 Intro to Operating Systems
CSCD 330 Network Programming
Programming with Shared Memory Specifying parallelism
CSE 332: Concurrency and Locks
A Sophomoric Introduction to Shared-Memory Parallelism and Concurrency Lecture 4 Shared-Memory Concurrency & Mutual Exclusion Dan Grossman Last Updated:
Presentation transcript:

A Sophomoric Introduction to Shared-Memory Parallelism and Concurrency Lecture 4 Shared-Memory Concurrency & Mutual Exclusion Dan Grossman Last Updated: May 2011 For more information, see

Toward sharing resources (memory) Have been studying parallel algorithms using fork-join –Lower span via parallel tasks Algorithms all had a very simple structure to avoid race conditions –Each thread had memory “only it accessed” Example: array sub-range –On fork, “loaned” some of its memory to “forkee” and did not access that memory again until after join on the “forkee” Strategy won’t work well when: –Memory accessed by threads is overlapping or unpredictable –Threads are doing independent tasks needing access to same resources (rather than implementing the same algorithm) 2Sophomoric Parallelism & Concurrency, Lecture 4

Concurrent Programming Concurrency: Correctly and efficiently managing access to shared resources from multiple possibly-simultaneous clients Requires coordination, particularly synchronization to avoid incorrect simultaneous access: make somebody block –join is not what we want –block until another thread is “done using what we need” not “completely done executing” Even correct concurrent applications are usually highly non-deterministic: how threads are scheduled affects what operations from other threads they see when –non-repeatability complicates testing and debugging 3Sophomoric Parallelism & Concurrency, Lecture 4

Examples Multiple threads: 1.Processing different bank-account operations –What if 2 threads change the same account at the same time? 2.Using a shared cache (e.g., hashtable) of recent files –What if 2 threads insert the same file at the same time? 3.Creating a pipeline (think assembly line) with a queue for handing work to next thread in sequence? –What if enqueuer and dequeuer adjust a circular array queue at the same time? 4Sophomoric Parallelism & Concurrency, Lecture 4

Why threads? Unlike parallelism, not about implementing algorithms faster But threads still useful for: Code structure for responsiveness –Example: Respond to GUI events in one thread while another thread is performing an expensive computation Processor utilization (mask I/O latency) –If 1 thread “goes to disk,” have something else to do Failure isolation –Convenient structure if want to interleave multiple tasks and don’t want an exception in one to stop the other 5Sophomoric Parallelism & Concurrency, Lecture 4

Sharing, again It is common in concurrent programs that: Different threads might access the same resources in an unpredictable order or even at about the same time Program correctness requires that simultaneous access be prevented using synchronization Simultaneous access is rare –Makes testing difficult –Must be much more disciplined when designing / implementing a concurrent program –Will discuss common idioms known to work 6Sophomoric Parallelism & Concurrency, Lecture 4

Canonical example Correct code in a single-threaded world 7Sophomoric Parallelism & Concurrency, Lecture 4 class BankAccount { private int balance = 0; int getBalance() { return balance; } void setBalance(int x) { balance = x; } void withdraw(int amount) { int b = getBalance(); if(amount > b) try {throw WithdrawTooLargeException;} setBalance(b – amount); } … // other operations like deposit, etc. }

Interleaving Suppose: –Thread T1 calls x.withdraw(100) –Thread T2 calls y.withdraw(100) If second call starts before first finishes, we say the calls interleave –Could happen even with one processor since a thread can be pre-empted at any point for time-slicing If x and y refer to different accounts, no problem –“You cook in your kitchen while I cook in mine” –But if x and y alias, possible trouble… 8Sophomoric Parallelism & Concurrency, Lecture 4

A bad interleaving Interleaved withdraw(100) calls on the same account –Assume initial balance == 150 9Sophomoric Parallelism & Concurrency, Lecture 4 int b = getBalance(); if(amount > b) try{throw …;} setBalance(b – amount); int b = getBalance(); if(amount > b) try{throw …;} setBalance(b – amount); Thread 1 Thread 2 Time “Lost withdraw” – unhappy bank

Incorrect “fix” It is tempting and almost always wrong to fix a bad interleaving by rearranging or repeating operations, such as: 10Sophomoric Parallelism & Concurrency, Lecture 4 void withdraw(int amount) { if(amount > getBalance()) try {throw WithdrawTooLargeException;} // maybe balance changed setBalance(getBalance() – amount); } This fixes nothing! Narrows the problem by one statement (Not even that since the compiler could turn it back into the old version because you didn’t indicate need to synchronize) And now a negative balance is possible – why?

Mutual exclusion The sane fix: At most one thread withdraws from account A at a time –Exclude other simultaneous operations on A too (e.g., deposit) Called mutual exclusion: One thread using a resource (here: an account) means another thread must wait –a.k.a. critical sections, which technically have other requirements Programmer must implement critical sections –“The compiler” has no idea what interleavings should or shouldn’t be allowed in your program –Buy you need language primitives to do it! 11Sophomoric Parallelism & Concurrency, Lecture 4

Wrong! Why can’t we implement our own mutual-exclusion protocol? –It’s technically possible under certain assumptions, but won’t work in real languages anyway 12Sophomoric Parallelism & Concurrency, Lecture 4 class BankAccount { private int balance = 0; private bool busy = false; void withdraw(int amount) { while(busy) { /* “spin-wait” */ } busy = true; int b = getBalance(); if(amount > b) try {throw WithdrawTooLargeException;} setBalance(b – amount); busy = false; } // deposit would spin on same boolean }

Still just moved the problem! 13Sophomoric Parallelism & Concurrency, Lecture 4 while(busy) { } busy = true; int b = getBalance(); if(amount > b) try {throw …;} setBalance(b – amount); while(busy) { } busy = true; int b = getBalance(); if(amount > b) try {throw …;} setBalance(b – amount); Thread 1 Thread 2 Time “Lost withdraw” – unhappy bank

What we need There are many ways out of this conundrum, but we need help from the language One basic solution: Locks, or critical sections –OpenMP implements locks as well as critical sections. They do similar things, but have subtle differences between them #pragma omp critical { // allows only one thread access at a time in the code block // code block must have one entrance and one exit } omp_lock_t myLock; –omp locks have to be initialized before use. –Can be locked with omp_set_nest_lock, and unlocked with omp_unset_nest_lock –Only one thread may hold the lock –Allows for exception handling and non structured jumps 14Sophomoric Parallelism & Concurrency, Lecture 4

Why that works An ADT with operations omp_ init_nest_lock, omp_destroy_nest_lock, omp_set_nest_lock, omp_unset_nest_lock The lock implementation ensures that given simultaneous acquires and/or releases, a correct thing will happen –Example: Two acquires: one will “win” and one will block The nested version of OpenMP Locks means that nested function calls where each nested function acquires the lock will work as expected. A thread will not deadlock trying to re- acquire a lock it already holds. 15Sophomoric Parallelism & Concurrency, Lecture 4

Why that works Omp critical section The critical section implementation ensures that statements enclosed within it are executed by a single thread at a time and so a correct thing will happen –Example: Two attempts to enter critical section: one will “win” and one will block 16Sophomoric Parallelism & Concurrency, Lecture 4

Almost-correct pseudocode 17Sophomoric Parallelism & Concurrency, Lecture 4 class BankAccount { private int balance = 0; private Lock lk = new Lock(); … void withdraw(int amount) { lk.acquire(); /* may block */ int b = getBalance(); if(amount > b) try {throw WithdrawTooLargeException;} setBalance(b – amount); lk.release(); } // deposit would also acquire/release lk }

Some mistakes A lock is a very primitive mechanism –Still up to you to use correctly to implement critical sections Incorrect: Use different locks for withdraw and deposit –Mutual exclusion works only when using same lock Poor performance: Use same lock for every bank account –No simultaneous operations on different accounts Incorrect: Forget to release a lock (blocks other threads forever!) –Previous slide is wrong because of the exception possibility! 18Sophomoric Parallelism & Concurrency, Lecture 4 if(amount > b) { lk.release(); // hard to remember! try {throw WithdrawTooLargeException;} }

Other operations If withdraw and deposit use the same lock, then simultaneous calls to these methods are properly synchronized But what about getBalance and setBalance ? –Assume they’re public, which may be reasonable If they don’t acquire the same lock, then a race between setBalance and withdraw could produce a wrong result If they do acquire the same lock, then withdraw would block forever because it tries to acquire a lock it already has 19Sophomoric Parallelism & Concurrency, Lecture 4

Re-acquiring locks? Can’t let outside world call setBalance1, not protected by locks Can’t have withdraw call setBalance2, because the locks would be nested We can use intricate re- entrant locking scheme or better yet re-structure the code. Nested locking is not recommended 20Sophomoric Parallelism & Concurrency, Lecture 4 int setBalance1(int x) { balance = x; } int setBalance2(int x) { lk.acquire(); balance = x; lk.release(); } void withdraw(int amount) { lk.acquire(); … setBalanceX(b – amount); lk.release(); }

This code is easier to lock You may provide a setBalance() method for external use, but do NOT call it for the withdraw() member function. Instead, protect the direct modification of the data as shown here – avoiding nested locks 21Sophomoric Parallelism & Concurrency, Lecture 4 int setBalance(int x) { lk.acquire(); balance = x; lk.release(); } void withdraw(int amount) { lk.acquire(); … balance = (b – amount); lk.release(); }

Critical Section version – that works 22Sophomoric Parallelism & Concurrency, Lecture 4 class BankAccount { private int balance; int getBalance() { int temp; #pragma omp critical { temp = balance; } return temp; } void withdraw(int amount) { { #pragma omp critical { if(balance >= amount) balance -= amount; } // Note: no need for setBalance method() }

OpenMP Critical Section method Works great & is simple to implement Can’t be used as nested function calls where each function sets a critical section. For example foo() contains critical section which calls bar(), while bar() also contains critical section. This will not be allowed by OpenMP runtime! Generally better to avoid nested calls such as foo,bar example above, and as also shown in bank account example foil 17 where withdraw() calls setBalance() Can’t be used with try/catch exception handling (due to multiple exists from code block) If try/catch is required, consider using scoped locking – example to follow. 23Sophomoric Parallelism & Concurrency, Lecture 4

Scoped Locking method Works great & is fairly simple to implement More complicated than critical section method Can be used as nested function calls where each function sets a critical section. –Generally it is still better to avoid nested calls such as foo,bar example above, and as also shown in bank account example foil 17 where withdraw() calls setBalance() Can be used with try/catch exception handling The lock is released when the object or function goes out of scope. 24Sophomoric Parallelism & Concurrency, Lecture 4

Scoped locking example constructor 25Sophomoric Parallelism & Concurrency, Lecture 4 class BankAccount { private int balance; omp_nest_lock_t _guard; BankAccount() { balance = 500.0; omp_init_nest_lock (&_guard); } … void deposit (int amount) { omp_guard my_guard (_guard); balance += amount; } … }

Scoped locking - withdraw, get, set 26Sophomoric Parallelism & Concurrency, Lecture 4 class BankAccount { … void withdraw(double amount) { omp_guard my_guard (_guard); if(balance >= amount) balance -= amount; } int getBalance() { omp_guard my_guard (_guard); return balance; } void setBalance(int x) { omp_guard my_guard (_guard); balance = x; }

Scoped locking – guard class 27Sophomoric Parallelism & Concurrency, Lecture 4 class omp_guard { public: omp_guard (omp_nest_lock_t &lock); void acquire (); void release (); ~omp_guard (); private: omp_nest_lock_t *lock_; // Disallow copies or assignment omp_guard (const omp_guard &); void operator = (const omp_guard &); … }

Scoped locking – guard class 28Sophomoric Parallelism & Concurrency, Lecture 4 class omp_guard { … omp_guard::omp_guard (omp_nest_lock_t &lock) : lock_ (&lock) { acquire (); } void omp_guard::acquire () { omp_set_nest_lock (lock_); } … }

Scoped locking – guard class 29Sophomoric Parallelism & Concurrency, Lecture 4 class omp_guard { … void omp_guard::release () { omp_unset_nest_lock (lock_); }