After studying this chapter, you should be able to: Define virtual memory; describe the hardware and control structures that support virtual memory; describe the various OS mechanisms used to implement virtual memory; describe the virtual memory management mechanisms in UNIX, Linux, and Windows 7.
Chapter Virtual Memory • Real memory – Main memory, the actual RAM, where a process executes • Virtual memory is a storage allocation scheme in which secondary memory can be addressed as though it were part of main memory – Size is limited by the amount of secondary memory available • Virtual address is the address assigned to a location in virtual memory Keys to Virtual Memory 1) Memory references are logical addresses that are dynamically translated into physical addresses at run time – A process may be swapped in and out of main memory, occupying different regions at different times during execution 2) A process may be broken up into pieces (pages or segments) that not need to be located contiguously in main memory Breakthrough in Memory Management • If both of those two characteristics are present, – then it is not necessary that all of the pages or all of the segments of a process be in main memory during execution • If the next instruction and the next data location are in memory then execution can proceed Execution of a Process • OS brings into main memory a few pieces of the program – Resident set: portion of process that is in main memory • Execution proceeds smoothly as long as all memory references are to locations that are in the resident set • An interrupt (memory access fault) is generated when an address is needed that is not in main memory Execution of a Process • OS places the process in a blocking state • Piece of process that contains the logical address is brought into main memory – OS issues a disk I/O Read request – Another process is dispatched to run while the disk I/O takes place – An interrupt is issued when disk I/O complete which causes OS to place the affected process in the Ready state Implications of this new strategy • More efficient processor utilization – More processes may be maintained in main memory because only load in some of the pieces of each process – More likely a process will be in the Ready state at any particular time • A process may be larger than main memory – This restriction in programming is lifted – OS automatically loads pieces of a process into main memory as required Thrashing • A condition in which the system spends most of its time swapping pieces rather than executing instructions • It happens when OS frequently throws out a piece just before it is used • To avoid this, OS tries to guess, based on recent history, which pieces are least likely to be used in the near future Principle of Locality • Program and data references within a process tend to cluster only a few pieces of a process will be needed over a short period of time • It is possible to make intelligent guesses about which pieces will be needed in the future, which avoids thrashing • This suggests that virtual memory may work efficiently Performance of Processes in VM Environment • During the lifetime of the process, references are confined to a subset of pages Support Needed for Virtual Memory • Hardware must support paging and segmentation • OS must be able to manage the movement of pages and/or segments between secondary memory and main memory 10 Optimal policy • Selects for replacement that page for which the time to the next reference is the longest • Results in the fewest number of page faults but it is impossible to have perfect knowledge of future events • Serves as a standard to judge real-world algorithms 37 Optimal Policy Example • The optimal policy produces three page faults after the frame allocation has been filled 38 Least Recently Used (LRU) • Replaces the page that has not been referenced for the longest time • By the principle of locality, this should be the page least likely to be referenced in the near future • Difficult to implement – One approach is to tag each page with the time of last reference – This requires a great deal of overhead 39 LRU Example • The LRU policy does nearly as well as the optimal policy – In this example, there are four page faults 40 First-in, first-out (FIFO) • Treats page frames allocated to a process as a circular buffer • Pages are removed in round-robin style – Simplest replacement policy to implement • Page that has been in memory the longest is replaced – But, these pages may be needed again very soon if it hasn’t truly fallen out of use 41 FIFO Example • The FIFO policy results in six page faults – Note that LRU recognizes that pages and are referenced more frequently than other pages, whereas FIFO does not 42 Clock Policy • Uses an additional bit called a “use bit” • When a page is first loaded in memory or referenced, the use bit is set to • When it is time to replace a page, the OS scans the set flipping all 1’s to • The first frame encountered with the use bit already set to is replaced • Similar to FIFO, except that, any frame with a use bit of is passed over by the algorithm 43 Clock Policy Example incoming page 727 44 Clock Policy Example An asterisk indicates that the corresponding use bit is equal to The arrow indicates the current position of the pointer • Note that the clock policy is adept at protecting frames and from replacement 45 Replacement Policy Comparison Two conflicting constraints: We would like to have a small page fault rate in order to run efficiently, We would like to keep a small frame allocation 46 Page Buffering • Replacing a modified page is more costly than unmodified page because the former must be written to secondary memory • Solution: page buffering + FIFO – Replaced page remains in memory (only the entry in the page table for this page is removed) and is added to the tail of one of two lists • Free page list if page has not been modified • Modified page list if it has 47 Page Buffering • The free page list is a list of page frames available for reading in pages • When a page is to be read in, the page frame at the head of the list is used, destroying the page that was there • The important aspect is that the page to be replaced remains in memory – If the process references that page, it is returned to the resident set of that process at little cost – In effect, the free and modified page lists act as a cache of pages 48 Cleaning Policy • A cleaning policy is concerned with determining when a modified page should be written out to secondary memory • Demand cleaning – A page is written out only when it has been selected for replacement – Minimizes page writes – A process that suffers a page fault may have to wait for two page transfers before it can be unblocked decrease processor utilization 49 Cleaning Policy • Precleaning – Pages are written out in batches (before they are needed) – Reduces the number of I/O operations – Pages written out may have been modified again before they are replaced waste of I/O operations with unnecessary cleaning operations 50 Cleaning Policy • Better approach: incorporates page buffering • Cleans only pages that are replaceable – decouples the cleaning and replacement operations • Pages in the modified list are periodically written out in batches and moved to the free list – significantly reduces the number of I/O operations and therefore the amount of disk access time • Pages in the free list are either reclaimed if referenced again or lost when its frame is assigned to another page 51 ... system that supports 231=2Gbytes virtual memory with 29=512-byte pages The number of entries in a page table can be as many as 222 • Most virtual memory schemes store page tables in virtual memory. .. that virtual memory may work efficiently Performance of Processes in VM Environment • During the lifetime of the process, references are confined to a subset of pages Support Needed for Virtual Memory. ..Keys to Virtual Memory 1) Memory references are logical addresses that are dynamically translated into physical addresses at run time – A process may be swapped in and out of main memory, occupying