1 / 44

Discussion Week 2

Explore the concepts of concurrency, process and thread levels, and the use of Pthreads, Java Threads, and Semaphores for implementing concurrency in multi-threaded code. Understand the challenges of race conditions, deadlock, and critical regions, and learn how to implement locks using semaphores. Experiment and compare the execution of multi-threaded code with and without concurrency primitives.

pierrea
Télécharger la présentation

Discussion Week 2

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Discussion Week 2 • TA: Kyle Dewey

  2. Overview • Concurrency • Process level • Thread level • MIPS - switch.s • Project #1

  3. Process Level • UNIX/Linux: fork() • Windows: CreateProcess()

  4. fork()/waitpid() Example

  5. while( true ) { fork(); }

  6. Threading Overview

  7. User-space Threads • OS does not know about them • Handle their own scheduling • If one blocks, all block • Cannot exploit SMP

  8. Blocking Example User Thread 1 User Thread 2 User Thread 3 Process 1 Process 2 OS

  9. Thread Standpoint User Thread 1 User Thread 2 User Thread 3 Process 1 Process 2 OS

  10. OS Standpoint Process 1 Process 2 OS

  11. Blocking • OS only sees a process • OS blocks the process, in turn blocking all user-space threads

  12. SMP • Processes have only a single thread • Without kernel assistance, this cannot be changed • Only one thread means only one CPU Process 1 Process 2 OS

  13. Kernel-Assisted • OS has knowledge of threads • OS schedules them • Act like individual processes sharing an address space

  14. General Pros/Cons • Kernel threads can exploit SMP • Kernel threads will not cause all threads to block • User-space threads are lightweight • Context switch is cheap • Likely far less code

  15. These are the concepts!

  16. Then implementation happened...

  17. Question: Do Pthreads threads run in user-space or are they kernel-assisted?

  18. Answer: Yes.

  19. Pthreads • Really just a standard with a number of possible implementations • Implementation can be kernel-assisted or in user-space • Most OSes are kernel-assisted

  20. Pthreads Example

  21. Java Threads • Again, merely a standard • Most implement as kernel-assisted threads

  22. Java Example

  23. Kernel Thread Implementation • OS can implement threads however it likes • Pthreads and Java are libraries built on top of the threading primitives provided by the OS

  24. Linux vs. Windows • Linux provides the clone() system call • Threads are actually processes • Windows provides CreateThread() • Referred to as “lightweight processes”

  25. NACHOS Threads • Kernel-assisted • Cannot currently handle interrupts or preemption correctly • Similar to MS-DOS...until project 2

  26. MS-DOS/NACHOS • One thread of execution • One process can run • OS is more like a large, complex software library

  27. Thread Primitives • Fork() - acts much like pthread_create • Yield() - gives up the CPU for any other available threads • Sleep() - like yield, but calling thread is blocked • Finish() - terminates calling thread

  28. For Project 1 • Fork() creates, but does not immediately start running, a new thread • Though there is no I/O, Sleep() can still be called to block on waiting for a critical region to clear

  29. NACHOS Threads

  30. Concurrency • Looks easy • Really hard to get right • Really hard • No seriously, borderline impossible

  31. Race Condition • Different results are possible based on different process/thread orderings • Ordering may be correct 99.999% of the time

  32. Deadlock • Two processes/threads wait for each other to do something • While they wait, they do not do whatever it is they are waiting for • Potential outcome of a race condition

  33. Critical Region • A point in code where the ordering matters • Almost always this is some state that is shared between processes/threads Server accept from port1 accept from port2 do something with both Client connect to server:port1 connect to server:port2 do something with both

  34. Fixing the Problem • Do not share state • Only share read-only state • Carefully regulate write access to shared state

  35. Regulation • A critical region can be manipulated by only one thread at a time • Need a way to enforce that at most one thread at any time point is in such a region

  36. Solving in Java • Java provides the synchronized keyword for blocks • Only one thread at a time may access a block marked with the synchronized keyword int x = 0; public synchronized void set( int y ) {x = y;} public int get() {return x;}

  37. Who cares about Java? • Many concurrency primitives work exactly like this, just with a little more work • One call upon entrance to critical region, another upon exit • The entrance and exit are implicit through blocks with Java

  38. Semaphores • Simply a shared integer • One call decrements, another increments • By convention, 0 is locked, and values > 0 are unlocked • Values < 0 mean the semaphore is not working!

  39. Semaphores • Increment/decrement are atomic - they are uninterruptible • The highest possible number it can hold is equal to the max number of callers to the region it protects

  40. Example int x = 0; Semaphore s; public void set( int y ) { s.decrement(); // wait/P/down x = y; s.increment(); } // signal/V/up public int get() {return x;}

  41. Project 1 Task 1 • Experiment according to instructions • Explain the execution of multithreaded code • Add semaphores and contrast the difference

  42. Project 1 Task 2 • Implement locks - essentially semaphores with a maximum of one caller at a time • Given all the semaphore code to look at • Hint hint it is a special case of a semaphore

  43. Project 1 Task 3 • Implement conditions • Require a correct Lock implementation • Allows a group of threads to synchronize on a given section of code • Can enforce that all must be at the same point of execution • Block until this is true

  44. Project 1 Task 4 • Identify and describe a race condition in a given section of code • Fix the race condition using semaphores • Fix it another way using locks and/or conditions

More Related