1 / 77

Solution to Race Condition: Mutual Exclusion and Synchronization

Solution to Race Condition: Mutual Exclusion and Synchronization. Problems with concurrent execution. Concurrent processes (or threads) often need to share data (maintained either in shared memory or files) and resources

ruana
Download Presentation

Solution to Race Condition: Mutual Exclusion and Synchronization

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Solution to Race Condition: Mutual Exclusion and Synchronization

  2. Problems with concurrent execution • Concurrent processes (or threads) often need to share data (maintained either in shared memory or files) and resources • If there is no controlled access to shared data, some processes will obtain an inconsistent view of this data • The action performed by concurrent processes will then depend on the order in which their execution is interleaved

  3. Example • Process P1 and P2 are running this same procedure and have access to the same variable “a” • Processes can be interrupted anywhere • If P1 is first interrupted after user input and P2 executes entirely • Then the character echoed by P1 will be the one read by P2 !! static char a; void echo() { read( a); write( a); }

  4. Race Conditions • Situations like this where processes access the same data concurrently and the outcome of execution depends on the particular order in which the access takes place is called a race condition • How must the processes coordinate (or synchronize) in order to guard against race conditions?

  5. The critical section problem • When a process executes code that manipulates shared data (or resource), we say that the process is in it’s critical section (CS) (for that shared data) • The execution of critical sections must be mutually exclusive: at any time, only one process is allowed to execute in its critical section (even with multiple CPUs) • Then each process must request the permission to enter it’s critical section (CS)

  6. The critical section problem • The section of code implementing this request is called the entry section • The critical section (CS) might be followed by an exit section • The remaining code is the remainder section (RS) • The critical section problem is to design a protocol that the processes can use so that their action will not depend on the order in which their execution is interleaved (possibly on many processors)

  7. Framework for analysis of solutions • Each process executes at nonzero speed but no assumption on the relative speed of n processes General structure of a process: … repeat entry section critical section exit section remainder section forever … • Many CPU may be present but memory hardware prevents simultaneous access to the same memory location • No assumption about order of interleaved execution • For solutions: we need to specify entry and exit sections

  8. Requirements for a valid solution to the critical section problem Mutual Exclusion • At any time, at most one process can be in its critical section (CS) Progress • Only processes that are not executing in their RS can participate in the decision of who will enter next CS. • This selection cannot be postponed indefinitely Bounded Waiting • After a process has made a request to enter it’s CS, there is a bound on the number of times that the other processes are allowed to enter their CS • otherwise the process will suffer from starvation • Of course also no deadlock (no cycles)

  9. Types of solutions Software • algorithms who’s correctness does not rely on any other assumptions Hardware • rely on some special machine instructions Operating System supported solutions • provide some functions and data structures to the programmer to implement a solution

  10. Software solutions: Algorithm 1 Process Pi: repeat while(turn!=i){}; CS turn:=j; RS forever An execution view of Algorithm 1 Process P0: Process P1: repeat repeat while(turn!=0){}; while(turn!=1){}; CS CS turn:=1; turn:=0; RS RS forever forever

  11. Software solutions: Algorithm 1 (cont.) • The shared variable turn is initialized (to 0 or 1) before executing any Pi • Pi’s critical section is executed iff turn = i • Pi is busy waiting if Pj is in CS: mutual exclusion is satisfied • Progress requirement is not satisfied since it requires strict alternation of CSs • If a process requires its CS more often then the other, it cannot get it.

  12. Algorithm 2 Process Pi: repeat flag[i]:=true; while(flag[j]){}; CS flag[i]:=false; RS forever

  13. Keep 1 Bool variable for each process: flag[0] and flag[1] • Pi signals that it is ready to enter it’s CS by: flag[i]:=true • Mutual Exclusion is satisfied but not the progress requirement • If we have the sequence: T0: flag[0]:=true T1: flag[1]:=true Both process will wait forever to enter their CS: we have a deadlock

  14. Algorithm 3 (Peterson’s algorithm) Process Pi: repeat flag[i]:=true; // want in turn:=j; // let the other in while (flag[j]&turn=j){}; CS flag[i]:=false; // do not want in RS forever

  15. Initialization: flag[0]:=flag[1]:=false turn:= 0 or 1 • Willingness to enter CS specified by flag[i]:=true • If both processes attempt to enter their CS simultaneously, only one turn value will last • Exit section: specifies that Pi is unwilling to enter CS Execution view of Algorithm 3 Process P0: Process P1: repeat repeat flag[0]:=true; flag[1]:=true; turn:= 1; turn:= 0; while(flag[1]&turn=1){}; while(flag[0]&turn=0){}; CS CS flag[0]:=false; flag[1]:=false; RSRS forever forever

  16. Proof of correctness Mutual exclusion is preserved since: • P0 and P1 are both in CS only if flag[0] = flag[1] = true and only if turn = i for each Pi (impossible) The progress and bounded waiting requirements are satisfied: • Pi cannot enter CS only if stuck in while() with condition flag[ j] = true and turn = j. • If Pj is not ready to enter CS then flag[ j] = false and Pi can then enter its CS • If Pj has set flag[ j]=true and is in its while(), then either turn=i or turn=j • If turn=i, then Pi enters CS. If turn=j then Pj enters CS but will then reset flag[ j]=false on exit: allowing Pi to enter CS • but if Pj has time to reset flag[ j]=true, it must also set turn=i • since Pi does not change value of turn while stuck in while(), Pi will enter CS after at most one CS entry by Pj (bounded waiting)

  17. What about process failures? • If all 3 criteria (ME, progress, bounded waiting) are satisfied, then a valid solution will provide robustness against failure of a process in its remainder section (RS) • since failure in RS is just like having an infinitely long RS • However, no valid solution can provide robustness against a process failing in its critical section (CS) • A process Pi that fails in its CS does not signal that fact to other processes: for them Pi is still in its CS

  18. n-process solution: Bakery algorithm • Before entering their CS, each Pi receives a number. Holder of smallest number enter CS (like in banks, bakeries, ice-cream stores...) • When Pi and Pj receives a number: • if i<j then Pi is served first, else Pj is served first • Pi resets its number to 0 in the exit section Notation: (a,b) < (c,d) if a < c or if a = c and b < d max(a0,...ak) is a number b such that b >= ai for i=0,..k

  19. Process Pi: repeat choosing[i]:=true; number[i]:=max(number[0]..number[n-1])+1; choosing[i]:=false; for j:=0 to n-1 do { while (choosing[j]) {}; while (number[j]!=0 and (number[j],j)<(number[i],i)){}; } CS number[i]:=0; RS forever

  20. Shared data: • choosing: array[0..n-1] of boolean; initialized to false • number: array[0..n-1] of integer;initialized to 0 Correctness relies on the following fact: • If Pi is in CS and Pk has already chosen its number[k]!= 0, then (number[i],i) < (number[k],k) • the proof is left out, a bit complicated!!!

  21. Peterson's solution int turn; int interested[2]; Process Pi: repeat other:=1-process; interested[process]:=TRUE; turn:=process; while (turn==proces && (interested[other]==TRUE){}; CS interested[process]:=FALSE; RS forever

  22. Shared turn and interested[2] allow two processes to share a CS. • Correctness: • p1 will not enter CS unless P2 is outside CS, vise versa. Thus processes cannot be blocking • turn take values of either 1 or 2, not both, if both processes are at the while statement • even if interested[1] and interested[2] were both TRUE at the same time, turn can be either 1 or 2. Thus only one of them can enter the CS • Disadvantages: • busy wait • starvation is possible because of priority, where one of the processes has higher priority. In this case priority inversion can be used, in which the waiting process is given the same priority as the one in CS, temporarily.

  23. Drawbacks of software solutions Processes that are requesting to enter in their critical section are busy waiting

  24. Hardware solutions: interrupt disabling/enabling Process Pi: repeat disable interrupts critical section enable interrupts remainder section forever

  25. On a uniprocessor: • mutual exclusion is preserved but efficiency of execution is degraded • The reason is that while in CS, we cannot interleave execution with other processes that are in RS On a multiprocessor: • mutual exclusion is not preserved • CS is now atomic but not mutually exclusive • Generally not an acceptable solution

  26. Hardware solutions: special machine instructions • Normally, access to a memory location excludes other access to that same location • Extension: designers have proposed machines instructions that perform 2 actions atomically (indivisible) on the same memory location (ex: reading and writing) • The execution of such an instruction is also mutually exclusive (even with multiple CPUs) • They can be used to provide mutual exclusion but need to be complemented to avoid starvation and deadlock)

  27. The test-and-set instruction C description of test-and-set implementation: test the flag i, if it is zero set it to 1 and return true, else return false. bool testset(int& i) { if (i==0) { i=1; return true; } else { return false; } }

  28. An algorithm that uses testset for Mutual Exclusion: Shared variable b is initialized to 0 Only the first Pi who sets b enter CS Process Pi: repeat repeat{} until testset(b); CS b:=0; RS forever

  29. Using an assembly tsl instruction Another implementation of test and set instruction, using an assembly tsl instruction which copies the content of the flag to a register and sets the flag to 1. enter-region; tsl register, flag //copy and set flag to 1 cmp register, #0 jnz enter_region //loop if flag was 1 ret //ok to enter CS leave-region: mov flag, #0 //store 0 in flag ret //allow others to enter CS

  30. Mutual exclusion is preserved: if Pi enter CS, the other Pj are busy waiting • Problem: still using busy waiting • When Pi exits CS, the selection of the Pj who will enter CS is arbitrary: no bounded waiting, hence starvation is possible

  31. Using xchg for mutual exclusion Processors (ex: Pentium) often provide an atomic xchg(a,b) instruction that swaps the content of a and b. Process Pi: //b is a shared variable repeat k:=1 repeat xchg(k,b) until k=0; //busy wait until b is 0 CS b:=0; RS forever

  32. Shared variable b is initialized to 0 • Each Pi has a local variable k • The only Pi that can enter CS is the one who finds b=0 • This Pi excludes all the other Pj by setting b to 1 • xchg(a,b) suffers from the same drawbacks as test-and-set: busy wait and starvation possiblity

  33. Semaphores • Synchronization tool (provided by the OS) that do not require busy waiting • A semaphore S is an integer variable that, apart from initialization, can only be accessed through 2 atomic and mutually exclusive operations: • wait(S) • signal(S) • To avoid busy waiting: when a process has to wait, it will be put in a blocked queue of processes waiting for the same event

  34. Hence, in fact, a semaphore is a record (structure): type semaphore = record count: integer; queue: process list end; var S: semaphore; • When a process must wait for a semaphore S, it is blocked and put on the semaphore’s queue • The signal operation removes (acc. to a fair policy like FIFO) one process from the queue and puts it in the list of ready processes:

  35. Wait(S) and signal(S) operations wait(S): S.count--; if (S.count<0) { block this process place this process in S.queue } signal(S): S.count++; if (S.count<=0) { remove a process P from S.queue place this process P on ready list } s.count must be initialized to a nonnegative value (depending on application)

  36. Some observations on semaphore operations • When S.count >=0: the number of processes that can execute wait(S) without being blocked = S.count • When S.count<0: the number of processes waiting on S is = |S.count| • Atomicity and mutual exclusion: no 2 process can be in wait(S) and signal(S) (on the same S) at the same time (even with multiple CPUs) • Hence the blocks of code defining wait(S) and signal(S) are, in fact, critical sections • The critical sections defined by wait(S) and signal(S) are very short: typically 10 instructions

  37. Implementation of wait(S) and Signal(S): • uniprocessor: disable interrupts during these operations (ie: for a very short period). This does not work on a multiprocessor machine. • multiprocessor: use previous software or hardware schemes. The amount of busy waiting should be small.

  38. Using semaphores for solving critical section problems Process Pi: repeat wait(S); CS signal(S); RS forever For n processes: Initialize S.count to 1 Then only 1 process is allowed into CS (mutual exclusion) To allow k processes into CS, we initialize S.count to k

  39. Using semaphores to synchronize 2 processes (P1, P2) • Proper synchronization is achieved by having in P1: S1; signal(synch); • And having in P2: wait(synch); S2; • Statement S1 in P1 needs to be performed before statement S2 in P2 • Then define a semaphore “synch” • Initialize synch to 0

  40. The producer/consumer problem • A producer(P) process produces information that is consumed by a consumer (C) process • a print program produces characters that are consumed by a printer • an assembler produces object modules that are consumed by a loader • We need a buffer to hold items that are produced and eventually consumed • This is a paradigm for cooperating processes

  41. P/C: unbounded buffer • We assume first an unbounded buffer consisting of a linear array of elements, as in the figure • in points to the next item to be produced • out points to the next item to be consumed

  42. We need a semaphore S to perform mutual exclusion on the buffer: only 1 process at a time can access the buffer • We need another semaphore N to synchronize producer and consumer on the number N (= in - out) of items in the buffer • An item can be consumed only after it has been created • The producer is free to add an item into the buffer at any time: it performs wait(S) before appending and signal(S) afterwards to prevent customer access • It also performs signal(N) after each append to increment N • The consumer must first do wait(N) to see if there is an item to consume and use wait(S)/signal(S) to access the buffer

  43. Solution of P/C: unbounded buffer case Initialization: S.count:=1; N.count:=0; in:=out:=0; Producer: Consumer: repeat repeat produce v; wait(N); wait(S); wait(S); append(v);w=take(); signal(S); signal(S); signal(N); consume(w); forever forever

  44. Remarks: • Putting signal(N) inside the CS of the producer (instead of outside) has no effect since the consumer must always wait for both semaphores before proceeding • The consumer must perform wait(N) before wait(S), otherwise deadlock occurs if consumer enter CS while the buffer is empty • Using semaphores requires extreme attention...

  45. P/C: finite circular buffer of size k • P can produce only if the buffer is not full, C can consume only if there is at least one item in the buffer • we need a semaphore S to have mutual exclusion on buffer access • we need a semaphore N to synchronize producer and consumer on the number of consumable items • we need a semaphore E to synchronize producer and consumer on the number of empty spaces

  46. Solution of P/C: finite circular buffer of size k Initialization: S.count:=1; in:=0; N.count:=0; out:=0; E.count:=k; Producer: Consumer: repeat repeat produce v; wait(N); wait(E); wait(S); wait(S); append(v); w=take(); signal(S); signal(S); signal(E) signal(N); consume(w); forever forever

  47. The Dining Philosophers Problem A classical synchronization problem • 5 philosophers who only eat and think • Each need to use 2 forks for eating • We have only 5 forks • Illustrates the difficulty of allocating resources among process without deadlock and starvation • Each philosopher is a process • One semaphore per fork: • fork: array[0..4] of semaphores • Initialization: fork[i].count:=1 for i:=0..4

  48. A first attempt: Process Pi: repeat think; wait(fork[i]); wait(fork[i+1 mod 5]); eat; signal(fork[i+1 mod 5]); signal(fork[i]); forever Deadlock if each philosopher start by picking his left fork!

  49. A Second attempt: admit only 4 philosophers at a time that tries to eat (interesting solution!) Process Pi: repeat think; wait(T); wait(fork[i]); wait(fork[i+1 mod 5]); eat; signal(fork[i+1 mod 5]); signal(fork[i]); signal(T); forever

More Related