CS 61C: Great Ideas in Computer Architecture Lecture 15: Caches, Part 2

Similar documents
CS 61C: Great Ideas in Computer Architecture Lecture 15: Caches, Part 2

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 2

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 2

Review. You Are Here! Agenda. Recap: Typical Memory Hierarchy. Recap: Components of a Computer 11/6/12

CS 61C: Great Ideas in Computer Architecture Caches Part 2

10/19/17. You Are Here! Review: Direct-Mapped Cache. Typical Memory Hierarchy

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 3

Caches Part 1. Instructor: Sören Schwertfeger. School of Information Science and Technology SIST

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 2

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 2

10/16/17. Outline. Outline. Typical Memory Hierarchy. Adding Cache to Computer. Key Cache Concepts

Agenda. Cache-Memory Consistency? (1/2) 7/14/2011. New-School Machine Structures (It s a bit more complicated!)

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 3

Agenda. Recap: Components of a Computer. Agenda. Recap: Cache Performance and Average Memory Access Time (AMAT) Recap: Typical Memory Hierarchy

CS 61C: Great Ideas in Computer Architecture Lecture 16: Caches, Part 3

CS 61C: Great Ideas in Computer Architecture. Direct Mapped Caches, Set Associative Caches, Cache Performance

CS 61C: Great Ideas in Computer Architecture Direct- Mapped Caches. Increasing distance from processor, decreasing speed.

CS 61C: Great Ideas in Computer Architecture. Cache Performance, Set Associative Caches

CS3350B Computer Architecture

Show Me the $... Performance And Caches

Mo Money, No Problems: Caches #2...

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 3

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 3

Direct-Mapped and Set Associative Caches. Instructor: Steven Ho

Review: Performance Latency vs. Throughput. Time (seconds/program) is performance measure Instructions Clock cycles Seconds.

10/11/17. New-School Machine Structures. Review: Single Cycle Instruction Timing. Review: Single-Cycle RISC-V RV32I Datapath. Components of a Computer

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 1

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 1

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part 1

EE 4683/5683: COMPUTER ARCHITECTURE

Course Administration

Page 1. Memory Hierarchies (Part 2)

CSF Improving Cache Performance. [Adapted from Computer Organization and Design, Patterson & Hennessy, 2005]

ECE7995 (6) Improving Cache Performance. [Adapted from Mary Jane Irwin s slides (PSU)]

EECS151/251A Spring 2018 Digital Design and Integrated Circuits. Instructors: John Wawrzynek and Nick Weaver. Lecture 19: Caches EE141

CS 61C: Great Ideas in Computer Architecture

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Caches Part I

CS61C : Machine Structures

ECE331: Hardware Organization and Design

CSE 431 Computer Architecture Fall Chapter 5A: Exploiting the Memory Hierarchy, Part 1

Components of a Computer

EEC 170 Computer Architecture Fall Improving Cache Performance. Administrative. Review: The Memory Hierarchy. Review: Principle of Locality

CS61C : Machine Structures

Memory Hierarchy. Maurizio Palesi. Maurizio Palesi 1

CMPT 300 Introduction to Operating Systems

CS61C : Machine Structures

Lecture 12. Memory Design & Caches, part 2. Christos Kozyrakis Stanford University

Caches and Memory Hierarchy: Review. UCSB CS240A, Winter 2016

Memory hierarchy review. ECE 154B Dmitri Strukov

CS 61C: Great Ideas in Computer Architecture (Machine Structures) Mul$- Level Caches

Memory Hierarchy. ENG3380 Computer Organization and Architecture Cache Memory Part II. Topics. References. Memory Hierarchy

CSF Cache Introduction. [Adapted from Computer Organization and Design, Patterson & Hennessy, 2005]

CENG 3420 Computer Organization and Design. Lecture 08: Cache Review. Bei Yu

CS 61C: Great Ideas in Computer Architecture. Multilevel Caches, Cache Questions

Caches and Memory Hierarchy: Review. UCSB CS240A, Fall 2017

ECE 2300 Digital Logic & Computer Organization. More Caches

Lecture 17 Introduction to Memory Hierarchies" Why it s important " Fundamental lesson(s)" Suggested reading:" (HP Chapter

Topics. Digital Systems Architecture EECE EECE Need More Cache?

Levels in memory hierarchy

COSC 6385 Computer Architecture. - Memory Hierarchies (I)

Kathryn Chan, Kevin Bi, Ryan Wong, Waylon Huang, Xinyu Sui

Modern Computer Architecture

CS 61C: Great Ideas in Computer Architecture. Direct Mapped Caches

CS 61C: Great Ideas in Computer Architecture (Machine Structures) More Cache: Set Associa0vity. Smart Phone. Today s Lecture. Core.

CS152 Computer Architecture and Engineering Lecture 17: Cache System

CENG 3420 Computer Organization and Design. Lecture 08: Memory - I. Bei Yu

ECE331: Hardware Organization and Design

UCB CS61C : Machine Structures

COSC3330 Computer Architecture Lecture 19. Cache

Memory. Lecture 22 CS301

Review: New-School Machine Structures. Review: Direct-Mapped Cache. TIO Dan s great cache mnemonic. Memory Access without Cache

Review : Pipelining. Memory Hierarchy

Administrivia. CMSC 411 Computer Systems Architecture Lecture 8 Basic Pipelining, cont., & Memory Hierarchy. SPEC92 benchmarks

CS 152 Computer Architecture and Engineering. Lecture 7 - Memory Hierarchy-II

Computer Architecture Spring 2016

Instructors: Randy H. Katz David A. PaGerson hgp://inst.eecs.berkeley.edu/~cs61c/fa10. 10/4/10 Fall Lecture #16. Agenda

Memory Hierarchy. Maurizio Palesi. Maurizio Palesi 1

CS 152 Computer Architecture and Engineering. Lecture 7 - Memory Hierarchy-II

Caches (Writing) P & H Chapter 5.2 3, 5.5. Hakim Weatherspoon CS 3410, Spring 2013 Computer Science Cornell University

14:332:331. Week 13 Basics of Cache

EEC 170 Computer Architecture Fall Cache Introduction Review. Review: The Memory Hierarchy. The Memory Hierarchy: Why Does it Work?

Agenda. CS 61C: Great Ideas in Computer Architecture. Virtual Memory II. Goals of Virtual Memory. Memory Hierarchy Requirements

Q3: Block Replacement. Replacement Algorithms. ECE473 Computer Architecture and Organization. Memory Hierarchy: Set Associative Cache

And in Review! ! Locality of reference is a Big Idea! 3. Load Word from 0x !

CS161 Design and Architecture of Computer Systems. Cache $$$$$

UCB CS61C : Machine Structures

Chapter 5. Large and Fast: Exploiting Memory Hierarchy

Chapter 5. Large and Fast: Exploiting Memory Hierarchy

COSC 6385 Computer Architecture - Memory Hierarchies (I)

UCB CS61C : Machine Structures

Caches II. CSE 351 Spring Instructor: Ruth Anderson

Caching Basics. Memory Hierarchies

Memory Hierarchy Motivation, Definitions, Four Questions about Memory Hierarchy

Pick a time window size w. In time span w, are there, Multiple References, to nearby addresses: Spatial Locality

COEN-4730 Computer Architecture Lecture 3 Review of Caches and Virtual Memory

Lecture 11 Cache. Peng Liu.

CS 136: Advanced Architecture. Review of Caches

CS61C : Machine Structures

CMPSC 311- Introduction to Systems Programming Module: Caching

CS152 Computer Architecture and Engineering CS252 Graduate Computer Architecture Spring Caches and the Memory Hierarchy

Transcription:

CS 61C: Great Ideas in Computer Architecture Lecture 15: Caches, Part 2 Instructor: Sagar Karandikar sagark@eecsberkeleyedu hbp://insteecsberkeleyedu/~cs61c 1

So/ware Parallel Requests Assigned to computer eg, Search Katz Parallel Threads Assigned to core eg, Lookup, Ads Parallel InstrucPons >1 instrucpon @ one Pme eg, 5 pipelined instrucpons Parallel Data >1 data item @ one Pme eg, Add of 4 pairs of words Hardware descrippons All gates @ one Pme Programming Languages You Are Here! Harness Parallelism & Achieve High Performance Hardware Warehouse Scale Computer Core Memory Input/Output InstrucPon Unit(s) Main Memory Computer (Cache) Core Core FuncPonal Unit(s) A 0 +B 0 A 1 +B 1 A 2 +B 2 A 3 +B 3 Smart Phone Today s Lecture Logic Gates 2

Caches Review Principle of Locality Temporal Locality and SpaPal Locality Hierarchy of Memories (speed/size/cost per bit) to Exploit Locality Cache copy of data in lower level of memory hierarchy Direct Mapped Caches Cache design choice: Write- Through vs Write- Back 3

Review: Adding Cache to Computer Processor Control Enable? Read/Write Memory Input Datapath PC Address Cache Program Bytes Registers Write Data ArithmePc & Logic Unit (ALU) Read Data Data Output Processor- Memory Interface I/O- Memory Interfaces 4

How do we use this thing? Nothing changes from the programmer s perspecpve: SPll issuing lw and sw instrucpons Everything we ll talk about is in the hardware: breaking down the address indexing into the cache choosing a block by checking tag extracpng bytes using the offset Why should a programmer care? Understanding cache parameters = fast programs 5

Review: Processor Address Fields used by Cache Controller Block Offset: Byte address within block #bits = log_2(block size in bytes) Index: Selects which index (set) we want # bits = log_2(number of sets) Tag: Remaining porpon of processor address processor address size - offset bits - index bits Processor Address (32- bits total) Tag Set Index Block offset 6

Working with a Direct Mapped Cache 16 Byte Direct Mapped Cache Index 00 01 10 11 Valid Tag Data Let s say our cache is what we see on the ler From the picture, we see that there are 4 indices (sets) Given that it is a 16 Byte cache and that there are 4 sets (the same as blocks in a DM cache), each block is 4 bytes Our processor can address 2^6 bytes of memory, so it has 6 bit addresses 7

Mapping an address to our cache Using the info from the previous slide 16 Byte Direct Mapped Cache 4 sets (sets, blocks, indices are all interchangeable in a direct mapped cache) 4 byte blocks Processor issues 6 bit addresses (ie has a 64 Byte address space) We can figure out how to break down the address Offset bits = log_2(block size) = log_2(4 byte blocks) = 2 Index bits = log_2(number of sets) = log_2(4 sets) = 2 Tag = 6-2 - 2 = 2 5 4 3 2 1 0 Tag Index Offset 8

Geometry of our DM Cache 16 Byte Direct Mapped Cache Index 00 01 10 11 Valid Tag Data 00 00 00 00 01 00 00 10 00 00 11 00 01 00 00 01 01 00 01 10 00 01 11 00 10 00 00 10 01 00 10 10 00 10 11 00 11 00 00 11 01 00 11 10 00 11 11 00 T I O Main Memory Recall that our processor can only address 64 B of memory in this example So, we ve drawn all of memory here: 16 rows, each 4 bytes wide We ve drawn memory in the same aspect rapo as the cache - in units of blocks rather than bytes Since each block is 4 bytes wide, the addresses you see end in 0b00 9

Direct Mapped Cache Hardware One word blocks, cache size = 1Ki words (or 4KiB) 31 30 13 12 11 2 1 0 Block offset Valid bit ensures something useful in cache for this index Hit Compare Tag with upper part of Address to see if a Hit Tag 20 10 Index Index 0 1 2 1021 1022 1023 Valid Tag 20 Data Comparator 32 Data Read data from cache instead of memory if a Hit 10

Cache Names for Each OrganizaPon Fully AssociaPve : Block can go anywhere First design in last lecture Note: No Index field, but 1 comparator/block Direct Mapped : Block goes one place Note: Only 1 comparator Number of sets = number blocks N- way Set AssociaPve : N places for a block Number of sets = number of blocks / N Fully AssociaPve: N = number of blocks in cache Direct Mapped: N = 1 11

Range of Set- AssociaPve Caches For a fixed- size cache, each increase by a factor of 2 in associapvity doubles the number of blocks per set (ie, the number of ways ) and halves the number of sets decreases the size of the index by 1 bit and increases the size of the tag by 1 bit More AssociaPvity (more ways) Tag Index Block offset Note: IBM persists in calling sets ways and ways sets They re wrong 12

Review: Write- Through vs Write- Back So far we handled reads What about writes? Write- Through: Simpler control logic More predictable Pming simplifies processor control logic Easier to make reliable, since memory always has copy of data Write- Back More complex control logic More variable Pming (0,1,2 memory accesses per cache access) Usually reduces write traffic Harder to make reliable, somepmes cache has only copy of data 13

Administrivia HW3 Out Proj2-2 Out Register your Project 3 Teams Yes, caches are hard, but don t worry: 15 discussions on caches 1 lab on caches (where you use a simulator) 1 more lecture on caches (arer today) Homework on caches Guerrilla secpon on Caches next week Guerrilla secpon today, 6-8pm in Woz, on MIPS CPU The good news: You re halfway done! 14

Example: Direct- Mapped $ with 4 Single- Word Lines, Worst- Case Reference String Consider the main memory address reference string 0 16 0 16 0 16 0 16 Start with an empty cache - all blocks inipally marked as not valid 0 miss 16 miss 0 miss 16 miss 01 16 00 0 01 16 00 Mem(0) 00 Mem(0) 01 Mem(16) 00 Mem(0) 000 0 miss 0 01 16 miss 16 00 0 miss 0 01 16 miss 16 01 Mem(16) 00 Mem(0) 01 Mem(16) 00 Mem(0) 8 requests, 8 misses Effect due to conflict misses - two memory locapons that map into the same cache block 15

AlternaPve Block Placement Schemes DM placement: mem block 12 in 8 block cache: only one cache block where mem block 12 can be found (12 modulo 8) = 4 SA placement: four sets x 2- ways (8 cache blocks), memory block 12 in set (12 mod 4) = 0; either element of the set FA placement: mem block 12 can appear in any cache blocks 16

Set Example: 2- Way Set AssociaPve $ Cache 0 1 Way 0 1 0 1 V (4 words = 2 sets x 2 ways per set) Tag Q: Is it there? Data Compare all the cache tags in the set to the high order 3 memory address bits to tell if the memory block is in the cache 000 0 xx 000 1 xx 001 0 xx 001 1 xx 010 0 xx 010 1 xx 011 0 xx 011 1 xx 100 0 xx 100 1 xx 101 0 xx 101 1 xx 110 0 xx 110 1 xx 111 0 xx 111 1 xx T I O Main Memory One word blocks Two low order bits define the byte in the word (32b words) Q: How do we find it? Use next 1 low order memory address bit to determine which cache set (ie, modulo the number of sets in the cache) 17

Example: 4- Word 2- Way SA $ Same Reference String Consider the main memory address reference string Start with an empty cache - all blocks inipally marked as not valid 0 16 0 16 0 16 0 16 Set 0 Set 1 0 miss 16 miss 0 hit 16 hit 000 Mem(0) 000 Mem(0) 000 Mem(0) 000 Mem(0) 010 Mem(16) 010 Mem(16) 010 Mem(16) Way 0 Way 1 Way 0 Way 1 8 requests, 2 misses Solves the conflict effect that occurred in a direct- mapped cache (to a degree) Because now two memory locapons that map into the same cache set can co- exist! But what if we did 0, 16, 32, 48, 0, 16, 32, 48,? 18

Different OrganizaPons of an Eight- Block Cache Total size of $ in blocks is equal to number of sets x associamvity For fixed $ size (and block size), increasing associapvity decreases number of sets while increasing number of elements per set With eight blocks, an 8- way set- associapve $ is same as a fully associapve $ 19

Recall: Direct Mapped Cache Hardware One word blocks, cache size = 1Ki words (or 4KiB) 31 30 13 12 11 2 1 0 Block offset Valid bit ensures something useful in cache for this index Hit Compare Tag with upper part of Address to see if a Hit Tag 20 10 Index Index 0 1 2 1021 1022 1023 Valid Tag 20 Data Comparator 32 Data Read data from cache instead of memory if a Hit 20

Four- Way Set- AssociaPve Cache 2 8 = 256 sets each with four ways (each with one block) 31 30 13 12 11 2 1 0 Byte offset Tag 22 Index 8 Index 0 1 2 253 254 255 V Tag Data 0 1 2 253 254 255 V Tag Data 0 1 2 253 254 255 V Tag Data 0 1 2 253 254 255 Way 0 Way 1 Way 2 Way 3 V Tag Data 32 Hit 4x1 select Data 21

Not the same!: MulPword- Block Hit Direct- Mapped Cache Four words/block, cache size = 1Ki words Tag 31 30 13 12 11 4 3 2 1 0 Byte offset 20 Index 8 Block offset Data Index Valid 0 1 2 253 254 255 Tag 20 Data 32 22

Range of Set- AssociaPve Caches For a fixed- size cache, each increase by a factor of two in associapvity doubles the number of blocks per set (ie, the number or ways) and halves the number of sets decreases the size of the index by 1 bit and increases the size of the tag by 1 bit Used for tag compare Selects the set Selects the byte(s) in the block Tag Index Byte offset Decreasing associapvity Direct mapped (only one way) Smaller tags, only a single comparator Increasing associapvity Fully associapve (only one set) Tag is all the bits except byte offset 23

Costs of Set- AssociaPve Caches N- way set- associapve cache costs N comparators (delay and area) MUX delay (set selecpon) before data is available Data available arer set selecpon (and Hit/Miss decision) DM $: block is available before the Hit/Miss decision In Set- AssociaPve, not possible to just assume a hit and conpnue and recover later if it was a miss When miss occurs, which way s block selected for replacement? Least Recently Used (LRU): one that has been unused the longest (principle of temporal locality) Must track when each way s block was used relapve to other blocks in the set For 2- way SA $, one bit per set set to 1 when a block is referenced; reset the other way s bit (ie, last used ) 24

Cache Replacement Policies Random Replacement Hardware randomly selects a cache evict Least- Recently Used Hardware keeps track of access history Replace the entry that has not been used for the longest Pme For 2- way set- associapve cache, need one bit for LRU replacement Example of a Simple Pseudo LRU ImplementaPon Assume 64 Fully AssociaPve entries Hardware replacement pointer points to one cache entry Whenever access is made to the entry the pointer points to: Move the pointer to the next entry Otherwise: do not move the pointer (example of not- most- recently used replacement policy) Replacement Pointer Entry 0 Entry 1 : Entry 63 25

Benefits of Set- AssociaPve Caches Choice of DM $ or SA $ depends on the cost of a miss versus the cost of implementapon Largest gains are in going from direct mapped to 2- way (20%+ reducpon in miss rate) 26

Clickers/Peer InstrucPon For a cache with constant total capacity, if we increase the number of ways by a factor of 2, which statement is false: A: The number of sets could be doubled B: The tag width could decrease C: The number of tags could stay the same D: The block size could be halved E: Tag width must increase 27

Break 28

Understanding Cache Misses: The 3Cs Compulsory (cold start or process migrapon, 1 st reference): First access to a block in memory impossible to avoid; small effect for long running programs SoluPon: increase block size (increases miss penalty; very large blocks could increase miss rate) Capacity: Cache cannot contain all blocks accessed by the program SoluPon: increase cache size (may increase access Pme) Conflict (collision): MulMple memory locamons mapped to the same cache locamon SoluMon 1: increase cache size SoluMon 2: increase associamvity (may increase access Mme) 29

How to Calculate 3C s using a Cache Simulator 1 Compulsory: set cache size to infinity and fully associapve, and count number of misses 2 Capacity: Change cache size from infinity, usually in powers of 2, and count misses for each reducpon in size 16 MB, 8 MB, 4 MB, 128 KB, 64 KB, 16 KB 3 Conflict: Change from fully associapve to n- way set associapve while counpng misses Fully associapve, 16- way, 8- way, 4- way, 2- way, 1- way 30

3Cs Analysis Three sources of misses (SPEC2000 integer and floapng- point benchmarks) Compulsory misses 0006%; not visible Capacity misses, funcpon of cache size Conflict porpon depends on associapvity and cache size 31

Rules for Determining Miss Type for a Given Access PaBern in 61C 1) Compulsory: A miss is compulsory if and only if it results from accessing the data for the first Pme If you have ever brought the data you are accessing into the cache before, it's not a compulsory miss, otherwise it is 2) Conflict: A conflict miss is a miss that's not compulsory and that would have been avoided if the cache was fully associapve Imagine you had a cache with the same parameters but fully- associapve (with LRU) If that would've avoided the miss, it's a conflict miss 3) Capacity: This is a miss that would not have happened with an infinitely large cache If your miss is not a compulsory or conflict miss, it's a capacity miss 32

Cache Measurement Terms Hit rate: fracpon of accesses that hit in the cache Miss rate: 1 Hit rate Miss penalty: Pme to replace a block from lower level in memory hierarchy to cache Hit Pme: Pme to access cache memory (including tag comparison) AbbreviaPon: $ = cache 33

Average Memory Access Time (AMAT) Average Memory Access Time (AMAT) is the average to access memory considering both hits and misses in the cache AMAT = Time for a hit + Miss rate x Miss penalty 34

Improving Cache Performance AMAT = Time for a hit + Miss rate x Miss penalty Reduce the Pme to hit in the cache Eg, Smaller cache Reduce the miss rate Eg, Bigger cache Reduce the miss penalty Eg, Use mulpple cache levels (next Pme) 35

Impact of Larger Cache on AMAT? 1) Reduces misses (what kind(s)?) 2) Longer Access Pme (Hit Pme): smaller is faster Increase in hit Pme will likely add another stage to the pipeline At some point, increase in hit Pme for a larger cache may overcome the improvement in hit rate, yielding a decrease in performance Computer architects expend considerable effort oppmizing organizapon of cache hierarchy big impact on performance and power! 36

Clickers: Impact of longer cache blocks on misses? For fixed total cache capacity and associapvity, what is effect of longer blocks on each type of miss: A: Decrease, B: Unchanged, C: Increase Compulsory? Capacity? Conflict? 37

Clickers: For fixed capacity and fixed block size, how does increasing associapvity affect AMAT? 38

Cache Design Space Several interacpng dimensions Cache size Block size AssociaPvity Replacement policy Write- through vs write- back Write allocapon OpPmal choice is a compromise Depends on access characterispcs Workload Use (I- cache, D- cache) Depends on technology / cost Simplicity oren wins Cache Size AssociaGvity Block Size Bad Good Factor A Factor B Less More 39

And, In Conclusion Name of the Game: Reduce AMAT Reduce Hit Time Reduce Miss Rate Reduce Miss Penalty Balance cache parameters (Capacity, associapvity, block size) 40