1 / 31

Operating Systems

Operating Systems. Practical Session 3, Scheduling. A quick recap. Quality criteria measures: Throughput – The number of completed processes per time unit. Turnaround time – The time interval between the process submission and its completion.

adila
Download Presentation

Operating Systems

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Operating Systems Practical Session 3, Scheduling

  2. A quick recap Quality criteria measures: • Throughput – The number of completed processes per time unit. • Turnaround time – The time interval between the process submission and its completion. • Waiting time – The sum of all time intervals in which the process was in the ready queue. • Response time – The time taken between submitting a command and the generation of first output. • CPU utilization – Percentage of time in which the CPU is not idle.

  3. A quick recap Two types of scheduling: • Preemptive schedulingA task may be rescheduled to operate at a later time (for example, it may be rescheduled by the scheduler upon the arrival of a “more important” task). • Non Preemptive scheduling (cooperative)Task switching can only be performed with explicitly defined system services (for example: the termination task, explicit call to yield() , I/O operation which changes the process state to blocking, etc’…).

  4. A quick recap Scheduling algorithms taught in class: • FCFS (First – Come, First – Served). • Non preemptive. • Convoy effect. • SJF (Shortest Job First). • Provably minimal with respect to the minimal average turn around time. • No way of knowing the length of the next CPU burst. • Can approximate according to: Tn+1=tn+(1- )Tn • Preemptive (Shortest Remaining Time First) or non preemptive. • Round Robin. • When using large time slices it imitates FCFS. • When using time slices which are closer to context switch time, more CPU time is wasted on switches.

  5. A quick recap • Guaranteed scheduling. • Constantly calculates the ratio between how much time the process has had since its creation and how much CPU time it is entitled to. • Guarantees 1/n of CPU time per process / user. • Priority scheduling. • A generalization of SJF (How?). • Multi Level Queue scheduling. • Partition the ready queue. • Each partition employs its own scheduling scheme. • A process from a lower priority group may run only if there is no higher priority process. May cause starvation!

  6. A quick recap • Dynamic Multi Level scheduling. • Takes into account the time spent waiting (the notion of aging to prevent starvation). • Highest Response Ratio Next: • Feedback scheduling. • Demote processes running longer. • Combine with aging to prevent starvation. • Two Level scheduling. • Involves schedulers for Memory-CPU operations, and another scheduler for Memory-Disk operations.

  7. The Completely Fair Scheduler (Linux Kernel 2.63) • CFS tries to model an “ideal, precise multitasking CPU” – one that could run multiple processes simultaneously, giving each equal processing power. • Obviously, this is purely theoretical, so how can we model it? • We may not be able to have one CPU run things simultaneously, but we can measure how much runtime each task has had and try and ensure that everyone gets their fair share of time.

  8. The Completely Fair Scheduler(cont.) • This is held in the vruntime variable for each task, and is recorded at the nanosecond level. • A lower vruntime indicates that the task has had less time to compute, and therefore has more need of the processor. • CFS uses a Red-Black tree to store, sort, and schedule tasks.

  9. The CFS Tree & Scheduling Algo. • The key for each node is the vruntime of the corresponding task. • To pick the next task to run, simply take the leftmost node. • The task accounts for its time with the CPU by adding its execution time to the virtual runtime and is then inserted back into the tree if runnable http://www.ibm.com/developerworks/linux/library/l-completely-fair-scheduler/

  10. Priorities • CFS doesn't use priorities directly but instead uses them as a decay factor for the time a task is permitted to execute. • Lower-priority tasks have higher factors of decay, where higher-priority tasks have lower factors of decay. • Priority is inverse to its effect – a higher priority task will accumulate vruntime more slowly, since it needs more CPU time. • Likewise, a low-priority task will have its vruntime increase more quickly, causing it to be preempted earlier.

  11. CFS Group Scheduling • CFS introduced group scheduling in release 2.6.24, adding another level of fairness. • Tasks can be grouped together, such as by the user which owns them. CFS can then be applied to the group level as well as the individual task level. • Group scheduling is another way to bring fairness to scheduling, particularly in the face of tasks that spawn many other tasks. • Example: Two groups exist, group A (contains 10 tasks) and group B (contains 2 tasks). Each group will receive an equal 50% of CPU power.

  12. Warm up (1) • Why bother with multiprogramming? • Assume processes in a given system wait for I/O 60% of the time. • What is the approximate CPU utilization with one process running? • What is the approximate CPU utilization with three processes running?

  13. Warm up (1) • If a process is blocking on I/O 60% of the time, than there is only 40% of CPU utilization. • At a given moment, the probability that all three processes are blocking on I/O is (0.6)3. That means that the CPU utilization is (1-(0.6)3)=0.784, or roughly 78%.

  14. Warm up (2) • Assume a single CPU machine with a non preemptive scheduler, attempts to schedule n independent processes. How many possible schedules exist? • Answer: This is exactly like ordering a set of n different characters forming a word of length n. That is, there are n! different possible schedules.

  15. Round Robin The following list of processes require scheduling (each requires x Time Units, or TUs): • PA – 6 TU • PB – 3 TU • PC – 1 TU • PD – 7 TU If RR scheduling is used, what quanta size should be used to achieve minimal average turnaround time? (assume 0 cost for context switches)

  16. Round Robin • Quanta = 1: • Quanta = 2: • Quanta = 3: 10.75 TU • Quanta = 4: 11.5 TU • Quanta = 5: 12.25 TU • Quanta = 6: 10.5 TU • Quanta = 7: 10.5 TU (15+9+3+17)/4=11 TU (14+10+5+17)/4=11.5 TU

  17. Round Robin Turnaround time depends on the size of the time quantum used. Note that it does not necessarily improve as time quantum size increases!

  18. Non preemptive scheduling (Taken from Tanenbaum) Assume 5 different jobs arrive at a computer center, roughly at the same time (same clock tick). Their expected run times are 10, 6, 2, 4 and 8 TU. Their (externally determined) priorities are 3, 5, 2, 1 and 4 respectively. For each of the following scheduling algorithm, determine the mean process turnaround time. Ignore process switching overhead. All jobs are completely CPU bound. • Priority Scheduling (non-preemptive). [Higher number means higher priority] • First come first served (in order 10, 6, 2, 4, 8) (non-preemptive) • Shortest job first (non-preemptive) • Priority Scheduling: (6+14+24+26+30)/5=20 • FCFS:(10+16+18+22+30)/5=19.2 • SJF:(2+6+12+20+30)/5=14

  19. Preemptive dynamic priorities scheduling (Taken from Silberschatz, 5-9) Consider the following preemptive priority scheduling algorithm with dynamically changing priorities: When a process is waiting for the CPU (in the ready queue, but not running), its priority changes at rate α; when it is running, its priority changes at rate β. All processes are given a priority of 0 when they enter the ready queue. The parameters alpha and beta can be set.Higher priority processes take higher values. • What is the algorithm that results from β> α > 0? • What is the algorithm that results from α < β< 0? • Is there a starvation problem in 1? in 2? explain.

  20. Preemptive dynamic priorities scheduling • >>0. To get a better feeling of the problem, we will create an example:C, P1, P2, P3 arrive one after the other and last for 3 TU, =1 and =2(bold marks the running process): The resulting schedule is FCFS. Slightly more formal: If a process is running it must have the highest priority value. While it is running, it’s priority value increases at a rate greater than any other waiting process. As a result, it will continue it’s run until it completes (or waits on I/O, for example). All processes in the waiting queue, increase their priority at the same rate, hence the one which arrived earliest will have the highest priority once the CPU is available.

  21. Preemptive dynamic priorities scheduling •  <<0. We will use (almost) the same example problem as before, but this time =-2, =-1: The resulting schedule is LIFO. More formally: If a process is running it must have the highest priority value. While it is running, it’s priority value decreases at a much lower rate than any other waiting process. As a result, it will continue it’s run until it completes (or waits on I/O, for example), or a new process with priority 0 is introduced. As before, all processes in the waiting queue, decrease their priority at the same rate, hence the one which arrived later will have the highest priority once the CPU is available.

  22. Preemptive dynamic priorities scheduling • In the first case it is easy to see that there is no starvation problem. When the kth process is introduced it will wait for at most (k-1)max{timei} Time Units. This number might be large but it is still finite. This is not true for the second case - consider the following scenario: P1 is introduced and receives CPU time. While still working a 2nd process, P2, is initiated. According to the scheduling algorithm in the second case, P2 will receive the CPU time and P1 will have to wait. As long as new processes will keep coming before P1 gets a chance to complete its run, P1 will never complete its task.

  23. Dynamic Multi Level scheduling An OS keeps two queues, Q1 and Q2. Each queue implements the round robin (RR) algorithm for all processes it holds. The OS prioritize processes in Q1, over those in Q2. When a process is created or returned from an I/O operation, it enters Q1. A process enters Q2 if it just finished running and it used up its whole time quantum. A process returning from I/O enters Q1 and has precedence over a process which did not start running. In our problem, we have the following processes: Process P1 – arrival time = 0, req.: 1 TU CPU, 1 TU IO, 3 TU CPU. Process P2 – arrival time = 2, req.: 2 CPU, 2 IO, 2 CPU. Process P3 – arrival time = 3, req.: 1 CPU, 3 IO, 3 CPU. Draw the Gantt table and compute the average TA and RT (turnaround and response time). Assume that the time quantum in Q1=1 TU, and the time quantum in Q2=2 TU. Further assume that the system has preemption. Computing the RT is based on the start of each I/O operation (you may think of the I/O as printing to stdout and that the user is waiting for this printout).

  24. Dynamic Multi Level scheduling The Gantt table: Avg. TA: (7+11+9)/3=9 Avg. RT: (1+6+2)/3=3

  25. Guaranteed scheduling (Quiz ‘05) Three processes run on a preemptive (switching is done on every round), guaranteed scheduling (ties are broken with preference to the lower ID process) OS. They require the use of both the CPU and I/O: P1 – 1 TU CPU, 4 TU I/O, 2 TU CPU P2 – 1 CPU, 2 I/O, 2 CPU P3 – 2 CPU, 1 I/O, 2 CPU For questions 1-3 draw a Gantt chart and compute the avg. TA time: • Identical arrival time, I/O is done on separate devices. • Identical arrival time, I/O is done on the same device with FCFS scheduling. • Process arrival time is: P1 at 0, P2 at 1 and P3 at 2. I/O is done on separate devices. • Would Round Robin produce the same results? Explain your answer.

  26. Guaranteed scheduling 1. The Gantt table: Avg. TA: (7+8+10)/3=8.33

  27. Guaranteed scheduling 2. The Gantt table: Avg. TA: (7+9+11)/3=9

  28. Guaranteed scheduling 3. The Gantt table: Avg. TA: (7+7+8)/3=7.33

  29. Guaranteed scheduling 4. No, RR implements fairness only between processes waiting in ready state, and hence no process will receive two consecutive time slices while another process is waiting. This is in contrast to Guaranteed Scheduling, which takes into account all the time the process spent in the system.

  30. Multi-core Scheduling נניח כי עומד לרשותנו מחשב בעל שני מעבדים (C1, C2). בכל רגע נתון עובדים שני המעבדים אלא אם כן אין יותר עבודות ממתינות.למערכת מגיעים בו זמנית אוסף של 13 תהליכים משלושה טיפוסים כמפורט להלן: • תהליכי A – תהליכים קצרים שמסתיימים לאחר יחידת זמן אחת (תהליך יחיד) • תהליכי B – תהליכים ארוכים מעט יותר שמסתיימים לאחר שתי יחידות זמן (7 תהליכים) • תהליכי C – תהליכים שמסתיימים לאחר שלוש יחידות זמן (5 תהליכים). עבור שני האלגוריתמים הבאים חשבו: • מה יהיה ה-turnaround time הממוצע ואיזה אלגוריתם מוצלח יותר עפ"י מדד זה? • בכמה זמן CPU נעשה שימוש ואיזה אלגוריתם מוצלח יותר עפ"י מדד זה? • מהו משך הזמן הנדרש לסיום החישוב ואיזה אלגוריתם מוצלח יותר עפ"י מדד זה? • תהליכים מטיפוס A ו-B מופנים למעבד C1 שמפעיל SJF על תהליכים אלו. כל תהליכי C מופנים למעבד C2. במידה ומעבד מסוים סיים את עבודתו לפני השני הוא מטפל בתהליכים הנותרים עפ"י עיקרון SJF. • התהליכים מופנים למעבדים השונים עפ"י SJF.

  31. Multi-core Scheduling נצייר טבלת Gantt לשני המעבדים: עבור האלגוריתם הראשון: ועבור האלגוריתם השני: כעת ניתן לענות על השאלות בקלות: • לראשון – avg. TA=(64+45)/13=109/13=8.38לשני – avg. TA=(55+45)/13=100/13=7.69ולכן ברור כי SJF יהיה מוצלח יותר כאשר נשתמש באלגוריתם השני. (עדיף על פני Affinity). • ברור שבשתי השיטות משך החישוב הנדרש יהיה זהה: 30. בשיטה אחת 15+15 ואילו בשיטה השניה 16+14. • במפתיע, למרות שהאלגוריתם השני מוצלח יותר מבחינתtime turnaround, הוא מסיים מאוחר יותר, כעבור 16 יחידות זמן, בעוד שהאלגוריתם הראשון מסיים לאחר 15 יחידות זמן.

More Related