Page Replacement Chap 21, 22. Dongkun Shin, SKKU

Similar documents
Page Replacement Chap 21, 22. Dongkun Shin, SKKU

Chapter 8: Virtual Memory. Operating System Concepts

OPERATING SYSTEM. Chapter 9: Virtual Memory

Chapter 9: Virtual Memory

CS307: Operating Systems

Chapter 9: Virtual-Memory

First-In-First-Out (FIFO) Algorithm

Chapter 9: Virtual Memory. Operating System Concepts 9 th Edition

Chapter 9: Virtual Memory. Operating System Concepts 9th Edition

Chapter 9: Virtual Memory

Chapter 10: Virtual Memory

Chapter 9: Virtual Memory. Operating System Concepts 9 th Edition

Operating System Concepts 9 th Edition

Chapter 9: Virtual Memory

Chapter 9: Virtual Memory

Chapter 9: Virtual Memory. Chapter 9: Virtual Memory. Objectives. Background. Virtual-address address Space

CS370 Operating Systems

Operating System Concepts

Chapter 8: Virtual Memory. Operating System Concepts Essentials 2 nd Edition

Virtual Memory Outline

Chapter 10: Virtual Memory. Background

Chapter 10: Virtual Memory. Background. Demand Paging. Valid-Invalid Bit. Virtual Memory That is Larger Than Physical Memory

Virtual Memory. Overview: Virtual Memory. Virtual address space of a process. Virtual Memory. Demand Paging

Where are we in the course?

Chapter 3: Virtual Memory ว ตถ ประสงค. Background สามารถอธ บายข อด ในการท ระบบใช ว ธ การจ ดการหน วยความจ าแบบเสม อนได

Virtual Memory COMPSCI 386

Chapter 9: Virtual Memory

CS370 Operating Systems

Optimal Algorithm. Replace page that will not be used for longest period of time Used for measuring how well your algorithm performs

Page Replacement Algorithms

Module 9: Virtual Memory

Module 9: Virtual Memory

PAGE REPLACEMENT. Operating Systems 2015 Spring by Euiseong Seo

Background. Demand Paging. valid-invalid bit. Tevfik Koşar. CSC Operating Systems Spring 2007

Virtual Memory. CSCI 315 Operating Systems Design Department of Computer Science

Virtual Memory. Overview: Virtual Memory. Virtual address space of a process. Virtual Memory

Virtual Memory. Virtual Memory. Demand Paging. valid-invalid bit. Virtual Memory Larger than Physical Memory

Chapters 9 & 10: Memory Management and Virtual Memory

Demand Paging. Valid-Invalid Bit. Steps in Handling a Page Fault. Page Fault. Transfer of a Paged Memory to Contiguous Disk Space

Memory Management. Virtual Memory. By : Kaushik Vaghani. Prepared By : Kaushik Vaghani

Background. Virtual Memory (2/2) Demand Paging Example. First-In-First-Out (FIFO) Algorithm. Page Replacement Algorithms. Performance of Demand Paging

Memory Management. To improve CPU utilization in a multiprogramming environment we need multiple programs in main memory at the same time.

Virtual Memory. CSCI 315 Operating Systems Design Department of Computer Science

CS6401- Operating System UNIT-III STORAGE MANAGEMENT

Chapter 9: Virtual-Memory Management. Operating System Concepts 8 th Edition,

Page Replacement. 3/9/07 CSE 30341: Operating Systems Principles

Lecture 17. Edited from slides for Operating System Concepts by Silberschatz, Galvin, Gagne

ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective. Part I: Operating system overview: Memory Management

Basic Memory Management

Chapter 8: Main Memory. Operating System Concepts 8th Edition

Principles of Operating Systems

Virtual Memory. Reading: Silberschatz chapter 10 Reading: Stallings. chapter 8 EEL 358

CSE 120. Translation Lookaside Buffer (TLB) Implemented in Hardware. July 18, Day 5 Memory. Instructor: Neil Rhodes. Software TLB Management

Virtual Memory III. Jo, Heeseung

!! What is virtual memory and when is it useful? !! What is demand paging? !! When should pages in memory be replaced?

CS420: Operating Systems

CS370 Operating Systems

Basic Page Replacement

Basic Memory Management. Basic Memory Management. Address Binding. Running a user program. Operating Systems 10/14/2018 CSC 256/456 1

VIRTUAL MEMORY READING: CHAPTER 9

Chapter 6: Demand Paging

Swapping. Operating Systems I. Swapping. Motivation. Paging Implementation. Demand Paging. Active processes use more physical memory than system has

instruction is 6 bytes, might span 2 pages 2 pages to handle from 2 pages to handle to Two major allocation schemes

Operating Systems. Overview Virtual memory part 2. Page replacement algorithms. Lecture 7 Memory management 3: Virtual memory

ADRIAN PERRIG & TORSTEN HOEFLER Networks and Operating Systems ( ) Chapter 6: Demand Paging

1. Background. 2. Demand Paging

Virtual Memory - Overview. Programmers View. Virtual Physical. Virtual Physical. Program has its own virtual memory space.

Topics: Virtual Memory (SGG, Chapter 09) CS 3733 Operating Systems

Memory management, part 2: outline

Memory management, part 2: outline. Operating Systems, 2017, Danny Hendler and Amnon Meisels

VII. Memory Management

CS 3733 Operating Systems:

Virtual Memory CHAPTER CHAPTER OBJECTIVES. 8.1 Background

Memory Management and Protection

UNIT - IV. What is virtual memory?

Even in those cases where the entire program is needed, it may not all be needed at the same time (such is the case with overlays, for example).

Virtual or Logical. Logical Addr. MMU (Memory Mgt. Unit) Physical. Addr. 1. (50 ns access)

Addresses in the source program are generally symbolic. A compiler will typically bind these symbolic addresses to re-locatable addresses.

MEMORY MANAGEMENT/1 CS 409, FALL 2013

Virtual Memory. ICS332 Operating Systems

Memory Management. Disclaimer: some slides are adopted from book authors slides with permission 1

Memory Management Virtual Memory

Chapter 8 & Chapter 9 Main Memory & Virtual Memory

Chapter 4 Memory Management

Memory Management. Disclaimer: some slides are adopted from book authors slides with permission 1

CS370 Operating Systems

CS370 Operating Systems

Virtual Memory Management

Operating Systems. Designed and Presented by Dr. Ayman Elshenawy Elsefy

Operating Systems Virtual Memory. Lecture 11 Michael O Boyle

Virtual Memory B: Objec5ves

Lecture 12: Demand Paging

Address spaces and memory management

Chapter 9: Virtual Memory

Chapter 8: Main Memory Chapter 9: Virtual Memory. Operating System Concepts 9th Edition

Memory Management. Chapter 4 Memory Management. Multiprogramming with Fixed Partitions. Ideally programmers want memory that is.

Virtual Memory. 1 Administrivia. Tom Kelliher, CS 240. May. 1, Announcements. Homework, toolboxes due Friday. Assignment.

CSE 4/521 Introduction to Operating Systems. Lecture 15 Virtual Memory I (Background, Demand Paging) Summer 2018

Memory Management Outline. Operating Systems. Motivation. Paging Implementation. Accessing Invalid Pages. Performance of Demand Paging

CS 333 Introduction to Operating Systems. Class 14 Page Replacement. Jonathan Walpole Computer Science Portland State University

Transcription:

Page Replacement Chap 21, 22 1

Virtual Memory Concept Virtual memory Concept A technique that allows the execution of processes that are not completely in memory Partition each user s program into multiple blocks Load into memory the blocks that is necessary at each time during execution Only part of the program needs to be in memory for execution Noncontiguous allocation Logical memory size is not constrained by the amount of physical memory that is available Separation of logical memory as perceived by users from physical memory 2

Virtual Memory Concept Virtual memory Benefits Easier programming Programmer no longer needs to worry about the amount of physical memory available Allows address spaces to be shared by several processes Higher multiprogramming degree Increase in CPU utilization and throughput (not in response time and turnaround time) Less I/O for loading and swapping processes into memory Faster execution of processes 3

Virtual Memory Concept Virtual memory Drawbacks Address mapping overhead Page fault handling overhead Not adequate for real-time (embedded) systems 4

Virtual Memory That is Larger Than Physical Memory OS make use of a larger, slower device to transparently provide the illusion of a large virtual address space 5

Memory Overlay (old system) Program-size > memory-size w/o OS support Requires support from compiler/linker/loader 6

Swapping A process can be swapped temporarily out of memory to a backing store (swap device) Process-level swapping Page-level swapping 7

Swapping Notes on swapping Time quantum vs swap time Time quantum should be substantially larger than swap time (context switch time) for efficient CPU utilization Memory areas to be swapped out Swap only what is actually used Pending I/O If the I/O is asynchronously accessing the user memory for I/O buffers, then the process cannot be swapped Solutions Never swap a process with pending I/O Execute I/O operations only into kernel buffers (and deliver it to the process memory when the process is swapped in) 8

Swapping Notes on swapping Swap device (swap file system) Swap space is allocated as a chunk of disk Contiguous allocation of the process image Fast access time Ordinary file system Discontiguous allocation of the file data blocks Focuses on the efficient file system space management Swap Partition vs. Swap File Linux kernel generates a map of disk blocks at swap-on time and from then on uses that map to perform swap I/O directly against the underlying disk queue, bypassing all caching, metadata and filesystem code. 9

Demand Paging Paging (Demand paging) system Partition the program into the same size blocks (pages) Loading of executable program Initially, load pages only as they are needed During execution, load the pages when they are demanded (referenced) Pages that are never accessed are never loaded into physical memory 10

Demand Paging 11

Demand Paging Address Mapping 12

Valid/Invalid (Present or Residence) Bit With each page table entry a valid/invalid bit is associated v: in-memory, memory resident i : not-in-memory Initially valid invalid bit is set to i on all entries Example of a page table snapshot: During MMU address translation, if valid invalid bit in page table entry is i page fault 13

Page Fault 14

Page Fault If there is a reference to a page, first reference to that page will trap to operating system: page fault 1. Operating system looks at another table to decide: Invalid reference abort Just not in memory 2. Find free frame 3. Swap page into frame via scheduled disk operation 4. Reset tables to indicate page now in memory Set validation bit = v 5. Restart the instruction that caused the page fault 15

Steps in Handling a Page Fault 16

Aspects of Demand Paging Actually, a given instruction could access multiple pages -> multiple page faults Consider fetch and decode of instruction which adds 2 numbers from memory and stores result back to memory Pain decreased because of locality of reference Hardware support needed for demand paging Page table with valid / invalid bit Secondary memory (swap device with swap space) Instruction restart ADD A,B,C Fetch and decode the instruction Fetch A Fetch B Add A and B Store the sum in C Page fault here! 17

Stages in Demand Paging 1. Trap to the operating system 2. Save the user registers and process state 3. Determine that the interrupt was a page fault 4. Check that the page reference was legal and determine the location of the page on the disk 5. Issue a read from the disk to a free frame: 1. Wait in a queue for this device until the read request is serviced 2. Wait for the device seek and/or latency time 3. Begin the transfer of the page to a free frame 6. While waiting, allocate the CPU to some other user 7. Receive an interrupt from the disk I/O subsystem (I/O completed) 8. Save the registers and process state for the other user 9. Determine that the interrupt was from the disk 10.Correct the page table and other tables to show page is now in memory 11.Wait for the CPU to be allocated to this process again 12.Restore the user registers, process state, and new page table, and then resume the interrupted instruction 18

Performance of Demand Paging Effective access time Memory access time 10 ~ 200 nanoseconds (Assume 200ns) Average paging service time: about 8 ms Page fault rate: p (0 p 1) EAT(Effective Access Time) EAT = (1-p)*ma + p*pagingtime = (1-p)*200 + p*8,000,000 = 200 + 7,999,800*p When p = 1/1000, EAT = 8.2 us (40 x ma) If we want less than 10% degradation, EAT = 200 + 7,999,800*p < 220 P < 0.0000025 (= 1/400,000) 19

Demand Paging Optimizations Swap space I/O faster than file system I/O even if on the same device Swap allocated in larger chunks, less management needed than file system Demand page in from program binary on disk, but discard rather than paging out when freeing frame Still need to write to swap space Pages not associated with a file (like stack and heap) anonymous memory Pages modified in memory but not yet written back to the file system Prefetching OS could guess that a page is about to be used, and thus bring it in ahead of time Mobile systems Typically don t support swapping Instead, demand page from file system and reclaim read-only pages (such as code) cf. zswap 20

Page Replacement Prevent over-allocation of memory Use modify (update, dirty) bit to reduce overhead of page transfers only modified pages are written to disk If modify == 1, the contents of the page in memory and in disk are not same Write-back (to disk) is necessary for the page Large virtual memory can be provided on a smaller physical memory 21

Page Replacement 1. Find the location of the desired page on disk 2. Find a free frame: - If there is a free frame, use it - If there is no free frame, use a page replacement algorithm to select a victim frame - Write victim frame to disk if dirty 3. Bring the desired page into the (newly) free frame; update the page and frame tables 4. Continue the process by restarting the instruction that caused the trap Note now potentially 2 page transfers for page fault increasing EAT 22

Page Replacement 23

Graph of Page Faults Versus The Number of Frames 24

When Replacements Really Occur? OS keeps a small portion of memory free more proactively Watermark scheme high watermark (HW) and low watermark (LW) When OS notices that there are fewer than LW pages available, a background thread (swap daemon or page daemon) that is responsible for freeing memory runs. The thread evicts pages until there are HW pages available. The background thread then goes to sleep. many systems will cluster or group a number of pages and write them out at once to the swap partition, thus increasing the efficiency of the disk 25

First-In-First-Out (FIFO) Algorithm Choose the page to be replaced based on when the page is previously loaded into memory Scheme Replace the oldest page Requirements Timestamping (memory load time for each page) is necessary Characteristics May replace frequently used pages FIFO anomaly (Belady s anomaly) In FIFO algorithm, page fault frequency may increase even if more memory frames are allocated 26

First-In-First-Out (FIFO) Algorithm 27

First-In-First-Out (FIFO) Algorithm FIFO algorithm: Example 4 page frames allocated, initially empty 28

First-In-First-Out (FIFO) Algorithm FIFO algorithm: Anomaly example 29

MIN algorithm (OPT algorithm) Proposed by Belady in 1966 Minimizes page fault frequency (proved) Scheme Replace the page that will not be used for the longest period of time Tie-breaking rule Unrealizable Page with greatest (or smallest) page number Can be used only when the process s reference string is known a priori Usage Performance measurement tool for replacement schemes 30

MIN algorithm (OPT algorithm) 31

MIN algorithm (OPT algorithm) MIN algorithm: Example 4 page frames allocated, initially empty 32

Least Recently Used (LRU) Algorithm Choose the page to be replaced based on the reference time Scheme Replace the page that has not been used for the longest period of time Requirements Timestamping (page reference time) is necessary Characteristics Based on program locality Approximates to the performance of MIN algorithm Used in most practical systems Drawbacks Timestamping overhead at every page reference Number of page faults increases steeply when the process executes large loop with insufficiently allocated memory 33

Least Recently Used (LRU) Algorithm 34

Least Recently Used (LRU) Algorithm LRU algorithm: Example 4 page frames allocated, initially empty 35

Implementation of LRU algorithm By counter Use PMT with count field Increment processor clock or counter for each memory access Record the value of processor clock or counter in the corresponding PMT entry for each page reference Can get the relative order of recent access to each page PMT search for selecting a page to be replaced 36

Implementation of LRU algorithm By stack Stack Stack for each process, whose entry is page number Maintains the stack elements (page numbers) in the order of recent access Can delete an element in the middle of the stack When no page fault Deletes the referenced page number from the stack, and inserts it on top of the stack When page fault Displaces the page whose number is at the bottom of the stack, deletes it from the stack, and inserts incoming page number on top of the stack 37

Implementation of LRU algorithm 38

Implementation of LRU algorithm 39

Implementation of LRU algorithm Example: LRU stack state change 4 page frames allocated, initially empty 40

LRU Approximation Algorithms LRU needs special hardware and still slow NUR(Not Used Recently) algorithm LRU approximation scheme Lower overhead than LRU algorithm Uses bit vectors Reference bit vector Update bit vector Scheme Check reference/update bit and choose a victim page Order of replacement (reference bit: r, update bit: m) 1 Replace the page with (r, m) = (0, 0) 2 Replace the page with (r, m) = (0, 1) 3 Replace the page with (r, m) = (1, 0) 4 Replace the page with (r, m) = (1, 1) 41

LRU Approximation Algorithms NUR algorithm 42

LRU Approximation Algorithms Additional reference-bits algorithm LRU approximation History register for each page Recording the reference bits at regular intervals Timer interrupts at regular intervals OS shifts the reference bit for each page into the high-order bit of its history register, shifting the other bits right by one bit and discarding the low-order bit Replacement based on the value of history register Interpret the value of the history register as unsigned integers Choose the page with smallest value as a victim 43

LRU Approximation Algorithms Additional reference-bits algorithm Example Value of the 8-bit history registers 00000000 No reference during 8 time periods 11111111 Referenced at least once in each period Page a with 01100100 and page b with 00110111 Page a has been used more recently that page b 44

LRU Approximation Algorithms Clock algorithm (Second-chance algorithm) Used for IBM VM/370 OS Uses reference bit No periodical reset for reference bits Choose the page to be replaced using pointer that circulates the list of pages (page frames) in memory 45

LRU Approximation Algorithms Clock algorithm Scheme Choose the page to be replaced with the pointer moving clockwise Mechanism 1 Check the reference bit r of the page that the pointer points 2 If r == 0, select the page as a victim 3 If r == 1, reset the reference bit to 0 and advances the pointer goto step-1 Characteristics Earlier memory load time higher probability of displacement Similar to FIFO algorithm Page replacement based on the reference bit Similar to LRU (or NUR) algorithm 46

LRU Approximation Algorithms Clock algorithm: Example 47

LRU Approximation Algorithms Enhanced clock algorithm Similar to clock algorithm Considers update (dirty, modified) bit as well as reference bit Scheme Choose the page to be replaced with the pointer moving clockwise Mechanism 1 Check (r, m) of the page that the pointer points 2 (0, 0): select the page as a victim, advances the pointer 3 (0, 1): set (r, m) to (0, 0), put the page on the cleaning list, goto step-6 4 (1, 0): set (r, m) to (0, 0), goto step-6 5 (1, 1): set (r, m) to (0, 1), goto step-6 6 Advances the pointer, goto step-1 48

LRU Approximation Algorithms Enhanced clock algorithm: Example 49

Counting Algorithms Keep a counter of the number of references that have been made to each page Not common Lease Frequently Used (LFU) Algorithm replaces page with smallest count Most Frequently Used (MFU) Algorithm based on the argument that the page with the smallest count was probably just brought in and has yet to be used 50

Workload Examples 51

Applications and Page Replacement All of these algorithms have OS guessing about future page access Some applications have better knowledge i.e. databases Memory intensive applications can cause double buffering OS keeps copy of page in memory as I/O buffer Application keeps page in memory for its own work Bypasses buffering, locking, etc O_DRIECT mode 52

Allocation of Frames How much space to allocate to each process Fixed allocation Fixed amount of memory is allocated to each process during entire execution of the process Variable allocation Variable amount of memory Proportional allocation Priority allocation Considerations It is necessary to estimate the space needs of the processes Too much allocation Memory waste Too small allocation Increase of page fault rate, performance degradation 53

Fixed Allocation Equal allocation For example, if there are 100 frames (after allocating frames for the OS) and 5 processes, give each process 20 frames Keep some as free frame buffer pool 54

Variable Allocation Proportional allocation Allocate according to the size of process Dynamic as degree of multiprogramming, process sizes change s a Priority allocation Use a proportional allocation scheme using priorities rather than size If process P i generates a page fault, i S s i size of process i m total number of allocation for p i p i frames si m S m 62 10 62 4 137 127 62 57 137 select for replacement one of its frames select for replacement a frame from a process with lower priority number s s a a 1 2 1 2 10 127 55

Global vs. Local Allocation Global replacement Allows a process to select a victim (replacement frame) from the set of all frames, even if that frame is currently allocated to another process one process can take a frame from another Process execution time can vary greatly Variable allocation based replacement VMIN(Variable MIN) algorithm WS(Working Set) algorithm PFF(Page Fault Frequency) algorithm Greater throughput so more common Local replacement Each process selects from only its own set of allocated frames More consistent per-process performance But possibly underutilized memory 56

Load control strategies Control the multiprogramming degree of a system Related to allocation strategy Should maintain multiprogramming degree at the appropriate range Underloaded System resource waste, performance degradation Overloaded System resource contention, performance degradation Thrashing (excessive paging activity) Plateau ranges are located at different positions for different systems 57

Thrashing 58

Thrashing If a process does not have enough pages, the page-fault rate is very high Page fault to get page Replace existing frame But quickly need replaced frame back This leads to: Low CPU utilization Operating system thinking that it needs to increase the degree of multiprogramming Another process added to the system Thrashing a process is busy swapping pages in and out 59

Demand Paging and Thrashing Why does demand paging work? Locality model Spatial locality and Temporal locality Process migrates from one locality to another Localities may overlap Why does thrashing occur? size of locality > total memory size Limit effects by using local or priority page replacement 60

Working set memory management Proposed by Denning in 1968 Working set Set of pages that the process intensively references at a time Set of pages in the most recent time interval of length Δ Working set of a process changes as time goes Definition : Working set The working set of process i at time t, W i (t, ) Set of pages that are referenced during the time interval [t-, t] by a process ( : window size, system parameter) 61

Working set memory management 62

Working set memory management Based on locality Let the working set to reside in memory every time Reduces page fault rate (reduces thrashing) Improve system performance D = W i, total demand frames if D > m Thrashing Policy if D > m, then suspend or swap out one of the processes Admission Control OOM killer Window size is an important system parameter in WS memory management Should consider the memory capacity and the characteristics of the processes in the system if too small will not encompass entire locality if too large will encompass several localities if = will encompass entire program 63

Working set memory management 64

Working set memory management 65

Working set memory management 66

Working set memory management 67

Working set memory management Characteristics A page can be displaced without incoming page No page may be displaced for incoming page Disadvantages Overhead Residence set, set of pages that reside in memory, should be updated for every page reference Residence set update independent of page faults 68

PFF (Page-Fault Frequency) Scheme More direct approach than WSS Residence set size page fault rate of the process Low page fault rate (long IFT (inter-fault time)) Reduces the number of page frames allocated to the process High page fault rate (short IFT) Allocates more page frames Residence set update, memory space adjustment Memory state changes only when page fault occurs Low overhead 69

PFF (Page-Fault Frequency) Scheme IFT > τ, Low Page fault rate IFT τ, High Page fault rate 70

Be Lazy: Demand Zeroing To add a page to your address space (heap) OS adds a page to your heap by finding a page in physical memory, zeroing it (required for security) and then mapping it into your address space (i.e., setting up the page table). High cost particularly if the page does not get used Demand zeroing OS puts an entry in the page table that marks the page inaccessible If the process then reads or writes the page, a trap takes place. When handling the trap, the OS notices that this is actually a demand-zero page. at this point, the OS then does the needed work of finding a physical page, zeroing it, and mapping it into the process s address space. If the process never accesses the page, all of this work is avoided 71

Be Lazy: Copy-on-Write Copy-on-Write When OS needs to copy a page from one address space to another, instead of copying it, it can map it into the target address space and mark it read-only in both address spaces. fast copy For read operation, no further action is taken For write operation, a trap occurs OS (lazily) allocate a new page, fill it with the data, and map this new page into the address space of the faulting process. fork() system call Creates a child process as a duplicate of its parent Should create a copy of the parent s address space for the child, duplicating the pages of the parent With copy-on-write scheme Allows the parent and the child processes initially share the pages (marked as copy-on-write) When any process writes to a shared page, a copy of the shared page is created 72

Be Lazy: Copy-on-Write After fork() system call After Parent modifies Page-C 73

Be Lazy: Copy-on-Write Notes on copy-on-write Only the copy-on-write pages that are modified by either process are copied All unmodified pages are shared by the parent and child processes Only the pages that can be modified can be marked as copyon-write Operating systems that use copy-on-write scheme Linux, Solaris, Windows XP 74

Be Lazy: Copy-on-Write Variation of fork() system call vfork() system call Let the child process use the address space of the parent Does not use copy-on-write Intended to be used when the child process calls exec() immediately after creation Should be used with caution Now obsolete 75

Memory-Mapped Files Memory-mapped file I/O allows file I/O to be treated as routine memory access by mapping a disk block to a page in memory A file is initially read using demand paging A page-sized portion of the file is read from the file system into a physical page Subsequent reads/writes to/from the file are treated as ordinary memory accesses Simplifies and speeds file access by driving file I/O through memory rather than read() and write() system calls Also allows several processes to map the same file allowing the pages in memory to be shared But when does written data make it to disk? Periodically and / or at file close() time For example, when the pager scans for dirty pages 76

Memory-Mapped File Technique for all I/O Some OSes uses memory mapped files for standard I/O Process can explicitly request memory mapping a file via mmap() system call Now file mapped into process address space Solaris: For standard I/O (open(), read(), write(), close()), mmap anyway But map file into kernel address space Process still does read() and write() Copies data to and from kernel space and user space Uses efficient memory management subsystem Avoids needing separate subsystem COW can be used for read/write non-shared pages Memory mapped files can be used for shared memory (although again via separate system calls) 77

MMAP Virtual Address 4G Physical Address HARD DISK Kernel Space MMU RAM fd=open( file ) 0xc0000000 read(8192byte) User Space 3G Read(2page) =8192bytes Page Cache offset Find 2 free pages in RAM & Read (512bytes x 16) Kernel Space read(8192byte) User Space Virtual Address 4G 0xc0000000 3G MMU mmap(2page) =8192bytes Physical Address RAM Page Cache HARD DISK fd=open( file ) offset Find 2 free pages in RAM & Read (512bytes x 16) 78

MMAP on Shared File Virtual Address in Process Physical Address 2 READ virt_addr1 Process 1 MMU Read(2page) =8192bytes RAM MMAP with MAP_SHARED flag HARD DISK 1WRITE (8192byte) Write data virt_addr2 Process 2 Virtual Address in Process Write(2page) =8192bytes Write data Page Cache Physical Address fd=open( file ) Write data offset Find 2 free pages in RAM & Read (512bytes x 16) 2 READ 1WRITE (2048byte) virt_addr1 virt_addr2! 3 2 1 Process 1 3 2 1 Process 2 MMU Read(1page) =4096bytes 1. copyon-write 2.Write(0.5page) =2048bytes RAM 3 2 1 Page Cache 2 HARD DISK fd=open( file ) offset MMAP with MAP_PRIVATE flag 79

Memory Mapped Files 80

Other Considerations -- Prepaging Prepaging (Prefetch) To reduce the large number of page faults that occurs at process startup Prepage all or some of the pages a process will need, before they are referenced But if prepaged pages are unused, I/O and memory was wasted Assume s pages are prepaged and α of the pages is used Is cost of s * α save pages faults > or < than the cost of prepaging s * (1- α) unnecessary pages? α near zero prepaging loses 81

Other Issues Page Size Sometimes OS designers have a choice Especially if running on custom-built CPU Page size selection must take into consideration: Fragmentation Page table size Resolution I/O overhead Number of page faults Locality TLB size and effectiveness Always power of 2, usually in the range 2 7 (128 bytes) to 2 22 (4MB) On average, growing over time Smaller page size + Smaller internal fragmentation + Match program locality more accurately o Reduction in total I/O o Better utilization of memory (Less total allocation of memory) - Larger page table size - Increase in I/O time - Increase in the number of page faults 82

Other Issues TLB Reach (Coverage) TLB Reach The amount of memory accessible from the TLB TLB Reach = (TLB Size) X (Page Size) Ideally, the working set of each process is stored in the TLB Otherwise there is a high degree of page faults Increase the number of entries in TLB Expensive Increase the Page Size lead to an increase in fragmentation as not all applications require a large page size Provide Multiple Page Sizes Allows applications that require larger page sizes the opportunity to use them without an increase in fragmentation Requires OS support One of the fields in TLB entry must indicate the size of the page (frame) corresponding to the TLB entry 83

Other Issues Program Structure Program restructuring System performance can be improved if the user (or compiler/linker/loader) has an awareness of the paged nature of memory or underlying demand paging system Program restructuring by user or compiler/linker/loader 84

Other Issues Program Structure Example Assumptions Paging system of 1KB page size sizeof(int): 4 Bytes // Program-1 int main() { int zar[256][256]; int i, j; for(j = 0; j < 256; j++) for(i = 0; i < 256; i++) zar[i][j] = 0; return 0; } For each execution of loop instance, pages 0~255 are referenced sequentially and this process is repeated 255 times For execution with no page faults, The process should have 256 page frames - Not practical 85

Other Issues Program Structure // Program-1 int main() { int zar[256][256]; int i, j; for(j = 0; j < 256; j++) for(i = 0; i < 256; i++) zar[i][j] = 0; return 0; } // Program-2 int main() { int zar[256][256]; int i, j; for(i = 0; i < 256; i++) for(j = 0; j < 256; j++) zar[i][j] = 0; return 0; } 86

Other Issues Program Structure Notes User Careful selection of data structures and programming structures can increase locality (hence lower page-fault rate) Stack: has good locality Hash table: produces bad locality Pointer: diminishes locality OOPs tend to have a poor locality of reference Linker/Loader Avoid placing routines across page boundaries Let routines that call each other many times be packed into the same page 87

Other Issues I/O interlock I/O Interlock Pages must sometimes be locked into memory Consider I/O - Pages that are used for copying a file from a device must be locked from being selected for eviction by a page replacement algorithm (lock bit) Another solution: never to execute I/O to user memory Pinning of pages to lock into memory 88