Operating Systems. Operating Systems Sina Meraji U of T

Similar documents
CS 153 Design of Operating Systems Winter 2016

Lecture 12: Demand Paging

CSE 120 Principles of Operating Systems

CSE 153 Design of Operating Systems

PAGE REPLACEMENT. Operating Systems 2015 Spring by Euiseong Seo

Virtual Memory III. Jo, Heeseung

All Paging Schemes Depend on Locality. VM Page Replacement. Paging. Demand Paging

Memory Management. Jinkyu Jeong Computer Systems Laboratory Sungkyunkwan University

Virtual Memory - II. Roadmap. Tevfik Koşar. CSE 421/521 - Operating Systems Fall Lecture - XVI. University at Buffalo.

Operating Systems Virtual Memory. Lecture 11 Michael O Boyle

CS 153 Design of Operating Systems Winter 2016

Reminder: Mechanics of address translation. Paged virtual memory. Reminder: Page Table Entries (PTEs) Demand paging. Page faults

CSE 120. Translation Lookaside Buffer (TLB) Implemented in Hardware. July 18, Day 5 Memory. Instructor: Neil Rhodes. Software TLB Management

CSE 120 Principles of Operating Systems

CSE 120 Principles of Operating Systems Spring 2017

VIRTUAL MEMORY II. Jo, Heeseung

How to create a process? What does process look like?

CS 333 Introduction to Operating Systems. Class 14 Page Replacement. Jonathan Walpole Computer Science Portland State University

CS162 Operating Systems and Systems Programming Lecture 11 Page Allocation and Replacement"

CS 333 Introduction to Operating Systems. Class 14 Page Replacement. Jonathan Walpole Computer Science Portland State University

Address Translation. Jinkyu Jeong Computer Systems Laboratory Sungkyunkwan University

EvicMng the best page #2: FIFO. #1: Belady s Algorithm. CS 537 Lecture 11 Page Replacement Policies 11/1/11. Michael SwiF

CS 153 Design of Operating Systems Spring 18

Virtual Memory Design and Implementation

Basic Page Replacement

Virtual or Logical. Logical Addr. MMU (Memory Mgt. Unit) Physical. Addr. 1. (50 ns access)

CS 318 Principles of Operating Systems

Virtual Memory Management

CISC 7310X. C08: Virtual Memory. Hui Chen Department of Computer & Information Science CUNY Brooklyn College. 3/22/2018 CUNY Brooklyn College

Virtual Memory. Today.! Virtual memory! Page replacement algorithms! Modeling page replacement algorithms

Operating Systems Lecture 6: Memory Management II

ECE7995 Caching and Prefetching Techniques in Computer Systems. Lecture 8: Buffer Cache in Main Memory (I)

VIRTUAL MEMORY READING: CHAPTER 9

Swapping. Jin-Soo Kim Computer Systems Laboratory Sungkyunkwan University

Swapping. Jinkyu Jeong Computer Systems Laboratory Sungkyunkwan University

Paging algorithms. CS 241 February 10, Copyright : University of Illinois CS 241 Staff 1

CS510 Operating System Foundations. Jonathan Walpole

Page Replacement Algorithms

3/3/2014! Anthony D. Joseph!!CS162! UCB Spring 2014!

Memory: Paging System

Memory Management. To improve CPU utilization in a multiprogramming environment we need multiple programs in main memory at the same time.

Page replacement algorithms OS

Virtual Memory. Chapter 8

Outline. 1 Paging. 2 Eviction policies. 3 Thrashing 1 / 28

Memory Management. Disclaimer: some slides are adopted from book authors slides with permission 1

Virtual Memory. Reading: Silberschatz chapter 10 Reading: Stallings. chapter 8 EEL 358

Last Class: Demand Paged Virtual Memory

Page Replacement. (and other virtual memory policies) Kevin Webb Swarthmore College March 27, 2018

!! What is virtual memory and when is it useful? !! What is demand paging? !! When should pages in memory be replaced?

Memory management, part 2: outline

Lecture 14 Page Replacement Policies

Chapter 8 Virtual Memory

Address spaces and memory management

Move back and forth between memory and disk. Memory Hierarchy. Two Classes. Don t

Memory Management Ch. 3

CS370 Operating Systems

MEMORY: SWAPPING. Shivaram Venkataraman CS 537, Spring 2019

Memory Management. Disclaimer: some slides are adopted from book authors slides with permission 1

Operating Systems (1DT020 & 1TT802) Lecture 9 Memory Management : Demand paging & page replacement. Léon Mugwaneza

Course Outline. Processes CPU Scheduling Synchronization & Deadlock Memory Management File Systems & I/O Distributed Systems

Operating Systems. Overview Virtual memory part 2. Page replacement algorithms. Lecture 7 Memory management 3: Virtual memory

ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective. Part I: Operating system overview: Memory Management

Chapter 8. Virtual Memory

a process may be swapped in and out of main memory such that it occupies different regions

CS450/550 Operating Systems

Memory management, part 2: outline. Operating Systems, 2017, Danny Hendler and Amnon Meisels

CS370 Operating Systems

Virtual Memory Outline

Chapter 8 Virtual Memory

Virtual Memory Design and Implementation

Chapter 4: Memory Management. Part 1: Mechanisms for Managing Memory

Chapter 8 Virtual Memory

Chapter 9: Virtual Memory

Virtual Memory. Virtual Memory. Demand Paging. valid-invalid bit. Virtual Memory Larger than Physical Memory

Lecture#16: VM, thrashing, Replacement, Cache state

ECE519 Advanced Operating Systems

Memory Management. Disclaimer: some slides are adopted from book authors slides with permission 1

Memory Management Virtual Memory

Virtual Memory. CSCI 315 Operating Systems Design Department of Computer Science

Week 2: Tiina Niklander

Memory Allocation. Copyright : University of Illinois CS 241 Staff 1

CS370 Operating Systems

Virtual Memory. CSCI 315 Operating Systems Design Department of Computer Science

Memory Management Prof. James L. Frankel Harvard University

Chapter 9: Virtual Memory

Operating Systems, Fall

ADRIAN PERRIG & TORSTEN HOEFLER Networks and Operating Systems ( ) Chapter 6: Demand Paging

Chapter 3 Memory Management: Virtual Memory

Memory Management. Chapter 4 Memory Management. Multiprogramming with Fixed Partitions. Ideally programmers want memory that is.

Page 1. Goals for Today" TLB organization" CS162 Operating Systems and Systems Programming Lecture 11. Page Allocation and Replacement"

Chapter 6: Demand Paging

Chapter 9: Virtual-Memory

CS 31: Intro to Systems Virtual Memory. Kevin Webb Swarthmore College November 15, 2018

Memory Management. Virtual Memory. By : Kaushik Vaghani. Prepared By : Kaushik Vaghani

Module 9: Virtual Memory

OPERATING SYSTEM. Chapter 9: Virtual Memory

Operating System Concepts

Chapter 8 & Chapter 9 Main Memory & Virtual Memory

MEMORY MANAGEMENT/1 CS 409, FALL 2013

Swapping. Operating Systems I. Swapping. Motivation. Paging Implementation. Demand Paging. Active processes use more physical memory than system has

Transcription:

Operating Systems Operating Systems Sina Meraji U of T

Recap Last time we looked at memory management techniques Fixed partitioning Dynamic partitioning Paging

Example Address Translation Suppose addresses are 16 bits, pages are 4K (4096 bytes) How many bits of the address do we need for offset? 12 bits (2^12 = 4096) What is the maximum number of pages for a process? 2^4 Bit 15 12 Pg # 11 Offset 0

Example Address Translation To translate virtual address: 0x3468 Extract page number (high-order 4 bits) -> page = vaddr >> 12 == 3 Get frame number from page table Combine frame number with page offset offset = vaddr % 4096 paddr = frame * 4096 + offset paddr = (frame << 12) offset Bit 15 12 11 0 0011 0100 0110 1000 Pg # Offset Logical address Page# Offset 3 468 Page# Frame# 0 1 1 4 2 5 3 7 7 468 Frame# Offset physical address

Page Lookups Overview Virtual Address Page number Offset Page table Physical Address Page frame Offset PTE What s wrong with this approach? Physical Memory Need 2 references for address lookup (first page table, then actual memory) Idea: Use hardware cache of page table entries Translation Lookaside Buffer (TLB) Small hardware cache of recently used translations

Page Lookups Overview Virtual Address Page number Offset Translation Lookaside Buffer TLB hit Physical Address PTE Page table Page frame Offset (in hardware cache) PTE TLB miss (in Memory) What s wrong with this approach? Physical Memory Need 2 references for address lookup (first page table, then actual memory) Idea: Use hardware cache of page table entries Translation Lookaside Buffer (TLB) Small, fully-associative hardware cache of recently used translations

TLBs Translate virtual page #s into PTEs (not physical addrs) Can be done in a single machine cycle TLBs implemented in hardware Fully associative cache (all entries looked up in parallel) Cache tags are virtual page numbers Cache values are PTEs (entries from page tables) With PTE + offset, can directly calculate physical address

TLBs TLBs are small (64 1024 entries) Still, address translations for most instructions are handled using the TLB >99% of translations, but there are misses (TLB miss) TLBs exploit locality Processes only use a handful of pages at a time 16-48 entries/pages (64-192K) Only need those pages to be mapped Hit rates are therefore very important

Managing TLBs Who places translations into the TLB (loads the TLB)? Hardware (Memory Management Unit) Software loaded TLB (OS)

Managing TLBs Who places translations into the TLB (loads the TLB)? Hardware (Memory Management Unit) Knows where page tables are in main memory OS maintains tables, HW accesses them directly Tables have to be in HW-defined format (inflexible) Software loaded TLB (OS) TLB faults to the OS, OS finds appropriate PTE, loads it in TLB Must be fast (but still 20-200 cycles) CPU ISA has instructions for manipulating TLB Tables can be in any format convenient for OS (flexible)

Managing TLBs (2) OS ensures that TLB and page tables are consistent When it changes the protection bits of a PTE, it needs to invalidate the PTE if it is in the TLB Reload TLB on a process context switch Invalidate all entries Why? When the TLB misses and a new PTE has to be loaded, a cached PTE must be evicted Choosing PTE to evict is called the TLB replacement policy Implemented in hardware, often simple

Summary so far: Paging Virtual Address Page number Offset Translation Lookaside Buffer PTE TLB hit Page table Physical Address Page frame Offset Page (in hardware cache) PTE Physical Memory TLB miss (in Memory) What happens if not all pages of all processes fit into physical memory?

Summary so far: Paging Virtual Address Page number Offset Translation Lookaside Buffer PTE TLB hit Page table Physical Address Page frame Offset Page (in hardware cache) PTE Physical Memory TLB miss (in Memory) What happens if page is evicted from main memory? Set PTE to invalid Store disk location in PTE Swap space Page Disk

How much space does a page table take up? Need one PTE per page 32 bit virtual address space w/ 4K pages = 2 20 PTEs 4 bytes/pte = 4MB/page table 25 processes = 100MB just for page tables! And modern processors have 64-bit address spaces -> 16 petabytes for page table! Solutions Hierarchical (multi-level) page tables Hashed page tables Inverted page tables

Managing Page Tables How can we reduce space overhead? Observation: Only need to map the portion of the address space actually being used (tiny fraction of entire addr space) How do we only map what is being used? Can dynamically extend page table Does not work if addr space is sparse (internal fragmentation) Use another level of indirection: two-level page tables (or multi-level page tables)

Motivation: two-level page tables Stack Page frame Virtual Address space Page Table 3 Heap Data Code How does address translation work now? Page frame Page Table 2 Page frame Page Table 1

Multilevel Page Tables (a) A 32-bit address with two page table fields. (b) Two-level page tables. Tanenbaum, Modern Operating Systems 3 e, (c) 2008 Prentice-Hall, Inc. All rights reserved. 0-13-6006639

Two-Level Page Tables Virtual addresses (VAs) have three parts: Master page number, secondary page number, and offset Master page table maps VAs to secondary page table Secondary page table maps page number to physical frame Offset selects address within physical frame Virtual Address Master page number Secondary Offset Physical Memory Page table Physical Address Page frame Offset Master Page Table Page frame Secondary Page Table

2-Level Paging Example 32-bit virtual address space 4K pages, 4 bytes/pte How many bits in offset? 4K = 12 bits, leaves 20 bits Want master/secondary page tables in 1 page each: 4K/4 bytes = 1K entries. How many bits to address 1K entries? 10 bits master = 10 bits offset = 12 bits secondary = 32 10 12 = 10 bits This is why 4K is common page size! Virtual Address Master page number Secondary Offset 10 bits 10 bits 12 bits

Pentium Address Translation To page table Data Cache/ Main Memory

Pentium Address Translation Data Cache/ Main Memory

Inverted Page Tables Keep one table with an entry for each physical page frame Entries record which virtual page # is stored in that frame Need to record process id as well Less space, but lookups are slower References use virtual addresses, table is indexed by physical addresses Use hashing (again!) to reduce the search time

Efficient Translations Our original page table scheme already doubled the cost of doing memory lookups One lookup into the page table, another to fetch the data Two-level page tables triple the cost! Two lookups into the page tables, a third to fetch the data And this assumes the page table is in memory TLB s hide the cost for frequently-used pages

Page allocation & eviction What happens when new page is allocated? Initially, pages are allocated from memory When memory fills up: Some other page needs to be evicted from memory This is why physical memory pages are called frames Evicted pages go to disk (the swap file) When it evicts a page, the OS sets the PTE as invalid and stores the location of the page in the swap file in the PTE

Recap: Paging Virtual Address Page number Offset Translation Lookaside Buffer PTE TLB hit Page table Physical Address Page frame Offset Page (in hardware cache) PTE Physical Memory TLB miss (in Memory) What happens if process accesses evicted page? Load page in memory Update PTE Swap space Page Disk

Page Faults What happens when a process accesses a page that has been evicted? 1. When a process accesses the page, the invalid PTE will cause a trap (page fault) 3. The trap will run the OS page fault handler 4. Handler uses the invalid PTE to locate page in swap file 5. Reads page into a physical frame, updates PTE to point to it 6. Restarts process

Policy Decisions Page tables, MMU, TLB, etc. are mechanisms that make virtual memory possible Next, we ll look at policies for virtual memory management: Fetch Policy when to fetch a page Placement Policy where to put the page Replacement Policy what page to evict to make room?

Demand Paging Application Load A Load B run stall stall page fault restart Operating System Memory Mgmt System phys. memory Disk Array Timing: Disk read is initiated when the process needs the page Request size: Process can only page fault on one page at a time, disk sees single page-sized read What alternative do we have?

Prepaging (aka Prefetching) Without Prepaging With Prepaging run run Load A stall Fetch A Load A stall Fetch A & B Load B stall Fetch B Load B Predict future page use at time of current fault On what should we base the prediction? What if it s wrong?

Policy Decisions Page tables, MMU, TLB, etc. are mechanisms that make virtual memory possible Next, we ll look at policies for virtual memory management: Fetch Policy when to fetch a page Demand paging vs. Prepaging Placement Policy where to put the page Are some physical pages preferable to others? Replacement Policy what page to evict to make room? Lots and lots of possible algorithms!

Placement Policy In paging systems, memory management hardware can translate any virtual-to-physical mapping equally well Why would we prefer some mappings over others? NUMA (non-uniform memory access) multiprocessors Any processor can access entire memory, but local memory is faster Cache performance Choose physical pages to minimize cache conflicts These are active research areas!

Policy Decisions Page tables, MMU, TLB, etc. are mechanisms that make virtual memory possible Next, we ll look at policies for virtual memory management: Fetch Policy when to fetch a page Demand paging vs. Prepaging Placement Policy where to put the page Are some physical pages preferable to others? Replacement Policy what page to evict to make room? Lots and lots of possible algorithms!

Evicting the best page The goal of the replacement algorithm is to reduce the fault rate by selecting the best victim page to remove Replacement algorithms are evaluated on a reference string by counting the number of page faults Let s start by cheating a little bit Assume we know the reference string what is the best replacement policy in this case?

Evicting the best page Page address list: 2,3,2,1,5,4,5,3,5,3,2 Cold misses: first access to a page (unavoidable) Capacity misses: caused by replacement due to limited size of memory 2 2 3 Lesson 1: 2 3 The best page to evict is the one never used again Will never fault on it 2 3 1 2 3 5

Evicting the best page Page address list: 2,3,2,1,5,4,5,3,5,3,2 Cold misses: first access to a page (unavoidable) Capacity misses: caused by replacement due to limited size of memory 2 2 3 Lesson 2: 2 3 Never is a long time, so picking the page closest to never is the next best thing Evicting the page that won t be used for the longest period of time minimizes the number of page faults Proved by Belady, 1966 2 3 1 2 3 5 4 3 5

Belady s Algorithm Belady s algorithm is known as the optimal page replacement algorithm because it has the lowest fault rate for any page reference stream (aka OPT or MIN) Idea: Replace the page that will not be used for the longest period of time Problem: Have to know the future perfectly Why is Belady s useful then? Use it as a yardstick Compare implementations of page replacement algorithms with the optimal to gauge room for improvement If optimal is not much better, then algorithm is pretty good If optimal is much better, then algorithm could use some work Random replacement is often the lower bound

What are possible replacement algorithms? First-in-first-out (FIFO) Least-recently-used (LRU) Least-frequently-used Most-frequently-used Many of these require book-keeping Let s start with algorithms that require only information contained in PTE

Page Table Entries (PTE) 1 1 1 3 26 M R V Prot Page Frame Number Modify (M) says whether or not page has been written Reference (R) says whether page has been accessed is cleared periodically (e.g. at clock interrupt) Valid (V) says whether PTE can be used Protection bits: what operations are allowed on page

Not-Recently-Used (NRU) Divide pages into 4 classes: Class1: Not referenced, not modified Class 2: Not referenced, modified Class 3: Referenced, not modified Class 4: Referenced, modified Remove page at random from lowestnumbered class that s not empty

First-In First-Out (FIFO) FIFO is an obvious algorithm and simple to implement Maintain a list of pages in order in which they were paged in On replacement, evict the one brought in longest time ago Why might this be good? Maybe the one brought in the longest ago is not being used Why might this be bad? Then again, maybe it s not We don t have any info to say one way or the other FIFO suffers from Belady s Anomaly The fault rate might actually increase when the algorithm is given more memory (very bad)

Example of Belady s anomaly Page Address List: 0,1,2,3,0,1,4,0,1,2,3,4 Youngest Oldest 0 0 1 1 0 2 2 1 0 3 3 2 1 0 0 3 2 1 1 0 3 4 4 1 0 0 4 1 0 1 4 1 0 2 2 4 1 3 3 2 4 4 3 2 4 3 frames, 9 faults Anomaly 0 1 2 3 0 1 4 0 1 2 3 4 Youngest 0 1 2 3 3 3 4 0 1 2 3 4 0 1 2 2 2 3 4 0 1 2 3 0 1 1 1 2 3 4 0 1 2 Oldest 0 0 0 1 2 3 4 0 1 4 frames, 10 faults

Second-Chance Idea: FIFO (First-in-first-out) considers only age NRU (Not recently used) considers only usage Maybe we should combine the two! Second chance algorithm: Don t evict the oldest page if it has been used. Evict the oldest page that has not been used. Pages that are used often enough to keep reference bits set will not be replaced

Implementing Second Chance (clock) Replace page that is old enough Arrange all of physical page frames in a big circle (clock) A clock hand is used to select a good LRU candidate Sweep through the pages in circular order like a clock If the ref bit (aka use bit) is off, it hasn t been used recently Evict the page If the ref bit is on Turn it off and go to next page Arm moves quickly when pages are needed Low overhead when plenty of memory Week 8 CSC369H1S 44

Modelling Clock 8 7 0 R=1 1 2 Physical Frames 6 3 5 4 1st page fault: Advance hand to frame 4, use frame 3 2nd page fault (assume none of these pages are referenced) Advance hand to frame 6, use frame 5

Least Recently Used (LRU) LRU uses reference information to make a more informed replacement decision Idea: We can t predict the future, but we can make a guess based upon past experience On replacement, evict the page that has not been used for the longest time in the past (Belady s: future) When does LRU do well? When does LRU do poorly? On average performs very well (close to Belady) But. CSC369H1S 11/03/2004

Implementing Exact LRU Option 1: Time stamp every reference Evict page with oldest time stamp Problems: Need to make PTE large enough to hold meaningful time stamp (may double size of page tables, TLBs) Need to examine every page on eviction to find one with oldest time stamp Option 2: Keep pages in a stack. On reference, move the page to the top of the stack. On eviction, replace page at bottom. Problems: Need costly software operation to manipulate stack on EVERY memory reference! CSC369H1S 11/03/2004

Modelling Exact LRU Page Address List: 0,1,2,3,0,1,4,0,1,2,3,4 3 Frames 0 1 2 3 0 1 4 0 1 2 3 4 MRU page 0 1 2 3 0 1 4 0 1 2 3 4 0 1 2 3 0 1 4 0 1 2 3 LRU page 0 1 2 3 0 1 4 0 1 2 4 Frames 0 1 2 3 0 1 4 0 1 2 3 4 MRU page 0 1 2 3 0 1 4 0 1 2 3 4 0 1 2 3 0 1 4 0 1 2 3 0 1 2 3 0 1 4 0 1 2 LRU page 0 1 2 3 3 3 4 0 1 10 faults 8 faults Week 8 CSC369H1S 48

Approximating LRU Exact LRU is too costly to implement LRU approximations use the PTE reference bit Basic Idea: Initially, all R bits are zero; as processes execute, bits are set to 1 for pages that are used Periodically examine the R bits we do not know order of use, but we know pages that were (or were not) used Additional-Reference-Bits Algorithm Keep a counter for each page At regular intervals, for every page do: Shift R bit into high bit of counter register Shift other bits to the right Pages with larger counters were used more recently Week 6 CSC369H1S 49

Counting-based Replacement Count number of uses of a page Least-Frequently-Used (LFU) Replace the page used least often Pages that are heavily used at one time tend to stick around even when not needed anymore Newly allocated pages haven t had a chance to be used much Most-Frequently-Used (MFU) Favours new pages Neither is common, both are poor approximations of OPT

What are possible replacement algorithms? Ease of implementation Performance Not-recently-used (NRU) First-in-first-out (FIFO) Least-recently-used (LRU) Least-frequently-used Most-frequently-used + ~ + ~ - ++ - - - -

What are possible replacement algorithms? Ease of implementation Performance Not-recently-used (NRU) First-in-first-out (FIFO) Least-recently-used (LRU) Least-frequently-used Most-frequently-used + ~ + ~ - ++ - - - -

What are possible replacement algorithms? Ease of implementation Performance Not-recently-used (NRU) First-in-first-out (FIFO) Least-recently-used (LRU) Least-frequently-used Most-frequently-used + ~ + ~ - ++ - - - - Second chance + +

Fixed vs. Variable Space In a multiprogramming system, we need a way to allocate memory to competing processes Problem: How to determine how much memory to give to each process? Fixed space algorithms Each process is given a limit of pages it can use When it reaches the limit, it replaces from its own pages Local replacement Some processes may do well while others suffer Variable space algorithms Process set of pages grows and shrinks dynamically Global replacement - one process can ruin it for the rest Local replacement - replacement and set size are separate

Working Set Model How do you decide how large the fixed or variable space for a process should be? Depends on access pattern Process 1 Process 2 6 1 5 2 1 6 2 7 5 1 4 4 3 3 4 1 3 4 4 4 5 pages {1,2,5,6,7} 2 or 3 pages {3,4} or {1,3,4}

Working Set Model A working set of a process is used to model the dynamic locality of its memory usage Defined by Peter Denning in 60 s Definition WS(t, ) = {pages P such that P was referenced in the time interval (t, t- )} t = time, = working set window (measured in page refs) A page is in the working set (WS) only if it was referenced in the last references 2 6 1 5 2 1 6 7 5 1 6 1 2 3 4 4 4 3 4 3 4 4 4 1 WS(t1) = {1,2,5,6,7} t1 WS(t2) = {3,4} t2

Working Set Size The working set size is the number of pages in the working set The number of pages referenced in the interval (t, t- ) The working set size changes with program locality During periods of poor locality, you reference more pages Within that period of time, the working set size is larger Intuitively, want the working set to be the set of pages a process needs in memory to prevent heavy faulting Each process has a parameter that determines a working set with few faults Denning: Don t run a process unless working set is in memory

Working Set Problems Problems How do we determine? How do we know when the working set changes? Too hard to answer So, working set is not used in practice as a page replacement algorithm However, it is still used as an abstraction The intuition is still valid When people ask, How much memory does Netscape need?, they are in effect asking for the size of Netscape s working set

Page Fault Frequency (PFF) Page Fault Frequency (PFF) is a variable space algorithm that uses a more ad-hoc approach Monitor the fault rate for each process If the fault rate is above a high threshold, give it more memory So that it faults less But not always (FIFO, Belady s Anomaly) If the fault rate is below a low threshold, take away memory Should fault more But not always Hard to use PFF to distinguish between changes in locality and changes in size of working set

Thrashing Page replacement algorithms avoid thrashing When more time is spent by the OS in paging data back and forth from disk than executing user programs No time spent doing useful work (making progress) In this situation, the system is overcommitted No idea which pages should be in memory to reduce faults Could just be that there isn t enough physical memory for all of the processes in the system Ex: Running Windows Vista with 4 MB of memory Possible solutions Swapping write out all pages of a process and suspend it Buy more memory

Windows XP Paging Policy Local page replacement Per-process FIFO Pages are stolen from processes using more than their minimum working set Processes start with a default of 50 pages XP monitors page fault rate and adjusts workingset size accordingly On page fault, cluster of pages around the missing page are brought into memory

Linux Paging Global replacement, like most Unix Modified second-chance clock algorithm Pages age with each pass of the clock hand Pages that are not used for a long time will eventually have a value of zero Continually under development