1 / 58

Chapter 10: Virtual Memory

Chapter 10: Virtual Memory. Background Demand Paging Process Creation Page Replacement Allocation of Frames Thrashing Operating System Examples. Background. Virtual memory – separation of user logical memory from physical memory.

ross
Télécharger la présentation

Chapter 10: Virtual Memory

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Chapter 10: Virtual Memory • Background • Demand Paging • Process Creation • Page Replacement • Allocation of Frames • Thrashing • Operating System Examples Operating System Concepts

  2. Background • Virtual memory – separation of user logical memory from physical memory. • Only part of the program needs to be in memory for execution. • Logical address space can therefore be much larger than physical address space. • Allows address spaces to be shared by several processes. • Allows for more efficient process creation. • Only parts of the program need be loaded into memory at the same time • Faster context switching, higher MPL (to a point) • Virtual memory can be implemented via: • Demand paging (more common, logical segment view, page implementation) • Demand segmentation ( more complex, segment replacement harder because of variable size) Operating System Concepts

  3. Virtual Memory That is Larger Than Physical Memory Operating System Concepts

  4. Demand Paging • Bring a page into memory only when it is needed. • Less I/O needed • Less memory needed • Faster response • More users(higher MPL hopefully, i.e. better CPU utilization, etc) • Page is needed  reference to it • invalid reference  abort • not-in-memory  bring to memory Operating System Concepts

  5. Transfer of a Paged Memory to Contiguous Disk Space Operating System Concepts

  6. Valid-Invalid Bit • With each page table entry a valid–invalid bit is associated(1  in-memory, 0 not-in-memory) • Initially valid–invalid but is set to 0 on all entries. • Example of a page table snapshot. • During address translation, if valid–invalid bit in page table entry is 0  page fault. Frame # valid-invalid bit 1 1 1 1 0  0 0 page table Operating System Concepts

  7. Page Table When Some Pages Are Not in Main Memory Operating System Concepts

  8. Page Fault • If there is ever a reference to a page, first reference will trap to OS  page fault • OS looks at another table to decide: • Invalid reference  abort. • Just not in memory. • Get empty frame. (from free frame list for example) • Swap page into frame. (schedule disk operation to move) • Reset tables, validation bit = 1. • Restart instruction: • Least Recently Used • block move • auto increment/decrement location Operating System Concepts

  9. Steps in Handling a Page Fault Operating System Concepts

  10. Restarting an instruction • ADD A, B, C (add A and B and place in C) • Fetch and decode instruction • Fetch A • Fetch B • Add A and B • Store sum in C • (page fault could happen during the store, so entire work will have to be redone) • Other examples that modify multiple locations while executing, thus a page fault can be problematic. Operating System Concepts

  11. What happens if there is no free frame? • Page replacement – find some page in memory, but not really in use, swap it out. • algorithm • performance – want an algorithm which will result in minimum number of page faults. • Same page may be brought into memory several times. Operating System Concepts

  12. Performance of Demand Paging • Page Fault Rate 0  p  1.0 • if p = 0 no page faults • if p = 1, every reference is a fault • Effective Access Time (EAT) EAT = (1-p)*ma + p*(page fault time) The page fault time is a function of many parameters: 1. Trap to the OS 2. Save the user registers and process state 3. Determine that the interrupt was a page fault 4. Check that the page reference was legal and determine the location of the page on the disk 5.1 Determine if free frame is available, if not, swap out LRU frame Operating System Concepts

  13. Performance of Demand Paging 5.2. Issue a read from the disk to a free frame a. wait in a queue for this device until read request is serviced b. wait for the device to seek and/or latency time c. begin the transfer of the page to the free frame 6. While waiting, allocate the CPU to some other process 7. Interrupt from the disk (I/O completed) 8. Save the registers and the process state for the other user (if applicable) 9. Determine that the interrupt was from the disk 10. Correct the page table and other tables to reflect that the desired page is now in memory 11. Wait for the CPU to be allocated to this process again 12.1 Restore the user registers, process state, and new page table 12.2 Resume execution from the interrupted instruction Not all steps are necessary in every case, some steps may have overlapping overhead times Operating System Concepts

  14. Performance of Demand Paging • Three major components of page-fault service time: • Service the page fault interrupt (relatively fast: microseconds) • Read in the page (relatively slow: milliseconds) • Restart the process (relatively fast: microseconds) • (remember that memory access time if VERY fast) (nanoseconds) • Suppose MA = 100 nanosec., PFST = 25 millisec. • EAT = (1-p)*ma + p*(page-fault service time) • EAT = (1-p)*100 + p*(25,000,000) • EAT = 24,999,900p + 100 (proportional to the page-fault rate) • Suppose you wanted to use demand paging only if there was no more than a 10% degradation in performance over not paging: • 100*(1.1) > 100 + 25,000,000p  p < .0000004  1 out of 2.5M Operating System Concepts

  15. Process Creation • Process can be quickly started using demand paging • Just page in in the page containing the first instruction • Demand paging can be use when reading a file from disk into memory, such files include binary executables • However, recall that process creation using syscall “fork()” can bypass initial paging altogether • Child process is a duplicate of the parent process • However, child might then call “exec()” to load it’s own image • Thus, copying parent address space may be unnecessary, thus COW • - Copy-on-Write (COW) • Initially share the same pages, • but is a write occurs, then copy page to child address space, • non-modified pages can be shared, reducing overhead • COW is used in Windows 2000, Linux, and Solaris Operating System Concepts

  16. Copy-on-Write • When COW is used, where do the free frames come from? • Typically from a “pool” of free frames • Uses a “zero-fill-on-demand” technique • Pages are zeroed-out just prior to being used (i.e. erased) • Several versions of UNIX provide a variation to “fork()” called “vfork()” for virtual memory fork: • Parent process suspended • Child has same address space as parent • Does not perform COW • Changes to data will be visible to parent when it resumes • Vfork() intended to be used when child will call exec() immediately Operating System Concepts

  17. Memory-Mapped Files • Memory-mapped file I/O allows file I/O to be treated as routine memory access by mapping a disk block to a page in memory. • A file is initially read using demand paging. A page-sized portion of the file is read from the file system into a physical page. Subsequent reads/writes to/from the file are treated as ordinary memory accesses. • Simplifies file access by treating file I/O through memory rather than read()write() system calls. • Also allows several processes to map the same file allowing the pages in memory to be shared. Operating System Concepts

  18. Memory Mapped Files Operating System Concepts

  19. Page Replacement • If we only page in the pages that are needed, we are tempted to increase the MPL. • This means that we over-allocate memory • If at some point a process wishes to use all its pages • Then we’re overbooked • Prevent over-allocation of memory by modifying page-fault service routine to include page replacement. • No free frames  page has to be replaced • Use modify (dirty) bit to reduce overhead of page transfers – only modified pages are written to disk. • Page replacement completes separation between logical memory and physical memory – large virtual memory can be provided on a smaller physical memory. Operating System Concepts

  20. Need For Page Replacement Operating System Concepts

  21. Basic Page Replacement • Find the location of the desired page on disk. • Find a free frame: - If there is a free frame, use it. - If there is no free frame, use a page replacement algorithm to select a victim frame. • Read the desired page into the (newly) free frame. Update the page and frame tables. • Restart the process. Operating System Concepts

  22. Page Replacement Operating System Concepts

  23. Page Replacement Algorithms • Want lowest page-fault rate. • Evaluate algorithm by running it on a particular string of memory references (reference string) and computing the number of page faults on that string. • In all our examples, the reference string is 1, 2, 3, 4, 1, 2, 5, 1, 2, 3, 4, 5. Operating System Concepts

  24. Graph of Page Faults Versus The Number of Frames Operating System Concepts

  25. First-In-First-Out (FIFO) Algorithm • Reference string: 1, 2, 3, 4, 1, 2, 5, 1, 2, 3, 4, 5 • 3 frames (3 pages can be in memory at a time per process) • 4 frames • FIFO Replacement – Belady’s Anomaly • more frames  less page faults 1 1 4 5 2 2 1 3 9 page faults 3 3 2 4 1 1 5 4 2 2 1 10 page faults 5 3 3 2 4 4 3 Operating System Concepts

  26. FIFO Page Replacement Operating System Concepts

  27. FIFO Illustrating Belady’s Anamoly Operating System Concepts

  28. Optimal Algorithm • Replace page that will not be used for longest period of time. • 4 frames example 1, 2, 3, 4, 1, 2, 5, 1, 2, 3, 4, 5 • How do you know this? • Used for measuring how well your algorithm performs. 1 4 2 6 page faults 3 4 5 Operating System Concepts

  29. Optimal Page Replacement Operating System Concepts

  30. Least Recently Used (LRU) Algorithm • Reference string: 1, 2, 3, 4, 1, 2, 5, 1, 2, 3, 4, 5 • Counter implementation • Every page entry has a counter; every time page is referenced through this entry, copy the clock into the counter. • When a page needs to be changed, look at the counters to determine which are to change. 1 5 2 3 5 4 4 3 Operating System Concepts

  31. LRU Page Replacement Operating System Concepts

  32. LRU Algorithm (Cont.) • Stack implementation – keep a stack of page numbers in a double link form: • Page referenced: • move it to the top • requires 6 pointers to be changed • No search for replacement Operating System Concepts

  33. Use Of A Stack to Record The Most Recent Page References Operating System Concepts

  34. LRU Approximation Algorithms • Hardware support is required in order for true LRU • Otherwise the overhead of doing it in software for each memory reference would slow the system intolerably. • Reference bit • With each page associate a bit, initially = 0 • When page is referenced bit set to 1. • Replace the one which is 0 (if one exists). We do not know the order, however. • More Bits (11000100 versus 01110111) (shift to right) • Second chance • Need reference bit. • Clock replacement. • If page to be replaced (in clock order) has reference bit = 1. then: • set reference bit 0. • leave page in memory. • replace next page (in clock order), subject to same rules. Operating System Concepts

  35. Second-Chance (clock) Page-Replacement Algorithm Operating System Concepts

  36. Counting Algorithms • Keep a counter of the number of references that have been made to each page. • LFU (Least frequently used) Algorithm: replaces page with smallest count.(suffers if the page it used initially a lot, but then it not used but still has a large count) (could shift the count bits to the right at regular intervals so that the count diminishes) • MFU Algorithm: based on the argument that the page with the smallest count was probably just brought in and has yet to be used. • Both of these are uncommon, are expensive to implement and are also not a good approximation to OPT. Operating System Concepts

  37. Allocation of Frames • Each process needs minimum number of pages. • Example: IBM 370 – 6 pages to handle SS MOVE instruction: • instruction is 6 bytes, might span 2 pages. • 2 pages to handle from. • 2 pages to handle to. • Number of levels of indirection of memory references is important, since if a page fault occurs, entire instruction must be started over, and all the frames must be there ultimately to support that instruction. • Two major allocation schemes. • fixed allocation • priority allocation Operating System Concepts

  38. Fixed Allocation • Equal allocation – e.g., if 100 frames and 5 processes, give each 20 pages. • Proportional allocation – Allocate according to the size of process. Operating System Concepts

  39. Priority Allocation • Use a proportional allocation scheme using priorities rather than size. • If process Pi generates a page fault, • select for replacement one of its frames. • select for replacement a frame from a process with lower priority number. Operating System Concepts

  40. Global vs. Local Allocation • Global replacement – process selects a replacement frame from the set of all frames; one process can take a frame from another. (i.e. start acquiring more total frames for itself) can • Can lead to wildly different execution times because the behavior depends on both it’s own page fault rate, but also that of other processes (the number of frames it has depends on the interacting behavior) • High priority can take frames from same class or lower • Number of frames allocated per process can change • Local replacement – each process selects from only its own set of allocated frames. • Number of frames allocated per process does not change • Suffers from the fact that it will not make less frequently used frames (from other processes) • Local replacement not typically used, since global tends to provide the best system throughput Operating System Concepts

  41. Thrashing • If a process does not have “enough” pages, the page-fault rate is very high. This leads to: • low CPU utilization. • operating system thinks that it needs to increase the degree of multiprogramming. • another process added to the system. • Thrashing a process is busy swapping pages in and out. • Minimum number of frames required by the architecture • Larger number are actively needed • If it doesn’t get the number it needs, it begins PF’ing quickly • When that happens, it must replace a page, and replace one that will be needed shortly • Bad cycle of page replacement begins Operating System Concepts

  42. Thrashing • Why does paging work?Locality model • Process migrates from one locality to another. • Localities may overlap. • Why does thrashing occur? size of locality > total memory size Operating System Concepts

  43. Locality In A Memory-Reference Pattern Operating System Concepts

  44. Working-Set Model •   working-set window  a fixed number of page references Example: 10,000 instruction • WSSi (working set of Process Pi) =total number of pages referenced in the most recent  (varies in time) • if  too small will not encompass entire locality. • if  too large will encompass several localities. • if  =   will encompass entire program. • D =  WSSi  total demand frames • if D > m  Thrashing • If D << m, another process can be initiated • Policy if D > m, then suspend one of the processes. Operating System Concepts

  45. Working-set model Operating System Concepts

  46. Keeping Track of the Working Set • Approximate with interval timer + a reference bit • Example:  = 10,000 • Timer interrupts after every 5000 time units. • Keep in memory 2 bits for each page. • Whenever a timer interrupts copy and sets the values of all reference bits to 0. • If one of the bits in memory = 1  page in working set. • Why is this not completely accurate? • b/c we don’t know exactly when the reference occurred • Could interrupt more often and maintain more history bits • But the overhead quickly builds and degrades performance • Improvement = 10 bits and interrupt every 1000 time units. • This strategy can work, but is perhaps clumsy, since the main objective it so prevent thrashing, so what can we do ? Operating System Concepts

  47. Page-Fault Frequency Scheme • Establish “acceptable” page-fault rate. • If actual rate too low, process loses frame. • If actual rate too high, process gains frame. Operating System Concepts

  48. Windows NT • Uses demand paging with clustering. Clustering brings in pages surrounding the faulting page. • Processes are assigned working set minimum and working set maximum. • Working set minimum is the minimum number of pages the process is guaranteed to have in memory. • A process may be assigned as many pages up to its working set maximum. • When the amount of free memory in the system falls below a threshold, automatic working set trimming is performed to restore the amount of free memory. • Working set trimming removes pages from processes that have pages in excess of their working set minimum. Operating System Concepts

  49. Other Considerations (Cont.) • I/O Interlock – Pages must sometimes be locked into memory. • Consider I/O. Pages that are used for copying a file from a device must be locked from being selected for eviction by a page replacement algorithm. Operating System Concepts

  50. Reason Why Frames Used For I/O Must Be In Memory Operating System Concepts

More Related