CS61C : Machine Structures

Similar documents
CS61C : Machine Structures

CS61C : Machine Structures

Lecture 7 More Memory Management Slab Allocator. Slab Allocator

CS61C : Machine Structures

CS61C : Machine Structures

CS61C Machine Structures. Lecture 8 - Introduction to the MIPS Processor and Assembly Language. 9/14/2007 John Wawrzynek

Review! Lecture 5 C Memory Management !

CS61C : Machine Structures

CS61C : Machine Structures

I ve been getting this a lot lately So, what are you teaching this term? Computer Organization. Do you mean, like keeping your computer in place?

Brought to you by CalLUG (UC Berkeley GNU/Linux User Group). Tuesday, September 20, 6-8 PM in 100 GPB.

Lecture #6 Intro MIPS; Load & Store Assembly Variables: Registers (1/4) Review. Unlike HLL like C or Java, assembly cannot use variables

Lectures 13 & 14. memory management

CS 61C: Great Ideas in Computer Architecture Intro to Assembly Language, MIPS Intro

CS 61C: Great Ideas in Computer Architecture Intro to Assembly Language, MIPS Intro

Review Pointers and arrays are virtually same C knows how to increment pointers C is an efficient language, with little protection

Lecture #6 Intro MIPS; Load & Store

CS 110 Computer Architecture Lecture 5: Intro to Assembly Language, MIPS Intro

CS 61C: Great Ideas in Computer Architecture Introduction to Assembly Language and RISC-V Instruction Set Architecture

CS61C Midterm Review on C & Memory Management

Another View of the Memory Hierarchy. Lecture #25 Virtual Memory I Memory Hierarchy Requirements. Memory Hierarchy Requirements

CS 61C: Great Ideas in Computer Architecture Introduction to Assembly Language and MIPS Instruction Set Architecture

A.Arpaci-Dusseau. Mapping from logical address space to physical address space. CS 537:Operating Systems lecture12.fm.2

Optimizing Dynamic Memory Management

CS61C : Machine Structures

Memory Allocation. Static Allocation. Dynamic Allocation. Dynamic Storage Allocation. CS 414: Operating Systems Spring 2008

UC Berkeley CS61C : Machine Structures

Outline. Lecture 40 Hardware Parallel Computing Thanks to John Lazarro for his CS152 slides inst.eecs.berkeley.

CS 430 Computer Architecture. C/Assembler Arithmetic and Memory Access William J. Taffe. David Patterson

Motivation for Dynamic Memory. Dynamic Memory Allocation. Stack Organization. Stack Discussion. Questions answered in this lecture:

CS61C : Machine Structures

CS61C : Machine Structures

CMSC 330: Organization of Programming Languages

Heap Management. Heap Allocation

Garbage Collection. Steven R. Bagley

Run-time Environments - 3

Run-time Environments -Part 3

ECE 598 Advanced Operating Systems Lecture 10

Processor. Lecture #2 Number Rep & Intro to C classic components of all computers Control Datapath Memory Input Output }

UC Berkeley CS61C : Machine Structures

Processor. Lecture #2 Number Rep & Intro to C classic components of all computers Control Datapath Memory Input Output

10.1. CS356 Unit 10. Memory Allocation & Heap Management

Memory Allocation. Copyright : University of Illinois CS 241 Staff 1

CS61C : Machine Structures

Discussion 4 Richard Guo SLAB, Buddy, Garbage Collection, and Intro to MIPS 01/30/09

Review. Disassembly is simple and starts by decoding opcode field. Lecture #13 Compiling, Assembly, Linking, Loader I

Acknowledgements These slides are based on Kathryn McKinley s slides on garbage collection as well as E Christopher Lewis s slides

The Art and Science of Memory Allocation

CS61C : Machine Structures

Dynamic Memory Allocation

xx.yyyy Lecture #11 Floating Point II Summary (single precision): Precision and Accuracy Fractional Powers of 2 Representation of Fractions

UC Berkeley CS61C : Machine Structures

CS61C : Machine Structures

CMSC 330: Organization of Programming Languages

UCB CS61C : Machine Structures

CS61C : Machine Structures

Lecture 33 Caches III What to do on a write hit? Block Size Tradeoff (1/3) Benefits of Larger Block Size

UC Berkeley CS61C : Machine Structures

Deallocation Mechanisms. User-controlled Deallocation. Automatic Garbage Collection

Lecture 13: Garbage Collection

15 Sharing Main Memory Segmentation and Paging

Lecture 3: Instruction Set Architecture

Heap Management portion of the store lives indefinitely until the program explicitly deletes it C++ and Java new Such objects are stored on a heap

Run-Time Environments/Garbage Collection

UCB CS61C : Machine Structures

Lecture 4: Memory Management & The Programming Interface

CS61C Machine Structures. Lecture 4 C Pointers and Arrays. 1/25/2006 John Wawrzynek. www-inst.eecs.berkeley.edu/~cs61c/

16 Sharing Main Memory Segmentation and Paging

CMSC 330: Organization of Programming Languages. Memory Management and Garbage Collection

Runtime. The optimized program is ready to run What sorts of facilities are available at runtime

G Programming Languages - Fall 2012

Speed. Lecture #39 Writing Really Fast Programs Example 1 bit counting Basic Idea

CS S-11 Memory Management 1

CS61C : Machine Structures

IEEE Standard 754 for Binary Floating-Point Arithmetic.

How about them A s!! Go Oaktown!! CS61C - Machine Structures. Lecture 4 C Structures Memory Management Dan Garcia.

One-Slide Summary. Lecture Outine. Automatic Memory Management #1. Why Automatic Memory Management? Garbage Collection.

CS61C : Machine Structures

Dynamic Memory Allocation. Basic Concepts. Computer Organization 4/3/2012. CSC252 - Spring The malloc Package. Kai Shen

Reference Counting. Reference counting: a way to know whether a record has other users

Inequalities in MIPS (2/4) Inequalities in MIPS (1/4) Inequalities in MIPS (4/4) Inequalities in MIPS (3/4) UCB CS61C : Machine Structures

CS 536 Introduction to Programming Languages and Compilers Charles N. Fischer Lecture 11

Names, Scope, and Bindings

CS 61C: Great Ideas in Computer Architecture Intro to Assembly Language, MIPS Intro. Machine Interpreta4on

UCB CS61C : Machine Structures

Numbers: positional notation. CS61C Machine Structures. Faux Midterm Review Jaein Jeong Cheng Tien Ee. www-inst.eecs.berkeley.

Lecture Notes on Garbage Collection

I-Format Instructions (3/4) Define fields of the following number of bits each: = 32 bits

Automatic Memory Management

Older geometric based addressing is called CHS for cylinder-head-sector. This triple value uniquely identifies every sector.

Class Information ANNOUCEMENTS

Robust Memory Management Schemes

Dynamic Memory Management

CS 241 Honors Memory

Dynamic Memory Management

Number Review. Lecture #3 More C intro, C Strings, Arrays, & Malloc Variables. Clarification about counting down

Chapter 10: Case Studies. So what happens in a real operating system?

Dynamic Memory Management! Goals of this Lecture!

CS61C : Machine Structures

CS577 Modern Language Processors. Spring 2018 Lecture Garbage Collection

Transcription:

inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures Lecture #5 Memory Management; Intro MIPS 2007-7-2 Scott Beamer, Instructor iphone Draws Crowds www.sfgate.com CS61C L5 Memory Management; Intro MIPS (1)

Review C has 3 pools of memory Static storage: global variable storage, basically permanent, entire program run The Stack: local variable storage, parameters, return address The Heap (dynamic storage): malloc() grabs space from here, free() returns it. Nothing to do with heap data structure! malloc() handles free space with freelist. Three different ways: First fit (find first one that s free) Next fit (same as first, start where ended) Best fit (finds most snug free space) One problem with all three is small fragments! CS61C L5 Memory Management; Intro MIPS (2)

Slab Allocator A different approach to memory management (used in GNU libc) Divide blocks in to large and small by picking an arbitrary threshold size. Blocks larger than this threshold are managed with a freelist (as before). For small blocks, allocate blocks in sizes that are powers of 2 e.g., if program wants to allocate 20 bytes, actually give it 32 bytes CS61C L5 Memory Management; Intro MIPS (3)

Slab Allocator Bookkeeping for small blocks is relatively easy: just use a bitmap for each range of blocks of the same size Allocating is easy and fast: compute the size of the block to allocate and find a free bit in the corresponding bitmap. Freeing is also easy and fast: figure out which slab the address belongs to and clear the corresponding bit. CS61C L5 Memory Management; Intro MIPS (4)

Slab Allocator 16 byte blocks: 32 byte blocks: 64 byte blocks: 16 byte block bitmap: 11011000 32 byte block bitmap: 0111 64 byte block bitmap: 00 CS61C L5 Memory Management; Intro MIPS (5)

Slab Allocator Tradeoffs Extremely fast for small blocks. Slower for large blocks But presumably the program will take more time to do something with a large block so the overhead is not as critical. Minimal space overhead No fragmentation (as we defined it before) for small blocks, but still have wasted space! CS61C L5 Memory Management; Intro MIPS (6)

Internal vs. External Fragmentation With the slab allocator, difference between requested size and next power of 2 is wasted e.g., if program wants to allocate 20 bytes and we give it a 32 byte block, 12 bytes are unused. We also refer to this as fragmentation, but call it internal fragmentation since the wasted space is actually within an allocated block. External fragmentation: wasted space between allocated blocks. CS61C L5 Memory Management; Intro MIPS (7)

Buddy System Yet another memory management technique (used in Linux kernel) Like GNU s slab allocator, but only allocate blocks in sizes that are powers of 2 (internal fragmentation is possible) Keep separate free lists for each size e.g., separate free lists for 16 byte, 32 byte, 64 byte blocks, etc. CS61C L5 Memory Management; Intro MIPS (8)

Buddy System If no free block of size n is available, find a block of size 2n and split it in to two blocks of size n When a block of size n is freed, if its neighbor of size n is also free, combine the blocks in to a single block of size 2n Buddy is block in other half larger block buddies NOT buddies Same speed advantages as slab allocator CS61C L5 Memory Management; Intro MIPS (9)

Allocation Schemes So which memory management scheme (K&R, slab, buddy) is best? There is no single best approach for every application. Different applications have different allocation / deallocation patterns. A scheme that works well for one application may work poorly for another application. CS61C L5 Memory Management; Intro MIPS (10)

Administrivia Third section is going to happen even though it isn t in telebears yet It will start meeting today Discussion MW 5-6 in 320 Soda Lab TuTh 5-7 in 271 Soda Attend your assigned section (if you go) If you are on the waitlist, the third section is your section HW3 and Proj1 will be posted in the next few days CS61C L5 Memory Management; Intro MIPS (11)

Automatic Memory Management Dynamically allocated memory is difficult to track why not track it automatically? If we can keep track of what memory is in use, we can reclaim everything else. Unreachable memory is called garbage, the process of reclaiming it is called garbage collection. So how do we track what is in use? CS61C L5 Memory Management; Intro MIPS (12)

Tracking Memory Usage Techniques depend heavily on the programming language and rely on help from the compiler. Start with all pointers in global variables and local variables (root set). Recursively examine dynamically allocated objects we see a pointer to. We can do this in constant space by reversing the pointers on the way down How do we recursively find pointers in dynamically allocated memory? CS61C L5 Memory Management; Intro MIPS (13)

Tracking Memory Usage Again, it depends heavily on the programming language and compiler. Could have only a single type of dynamically allocated object in memory E.g., simple Lisp/Scheme system with only cons cells (61A s Scheme not simple ) Could use a strongly typed language (e.g., Java) Don t allow conversion (casting) between arbitrary types. C/C++ are not strongly typed. Here are 3 schemes to collect garbage CS61C L5 Memory Management; Intro MIPS (14)

Scheme 1: Reference Counting For every chunk of dynamically allocated memory, keep a count of number of pointers that point to it. When the count reaches 0, reclaim. Simple assignment statements can result in a lot of work, since may update reference counts of many items CS61C L5 Memory Management; Intro MIPS (15)

Reference Counting Example For every chunk of dynamically allocated memory, keep a count of number of pointers that point to it. When the count reaches 0, reclaim. int *p1, *p2; p1 = malloc(sizeof(int)); p2 = malloc(sizeof(int)); *p1 = 10; *p2 = 20; p1 p2 Reference count = 1 20 Reference count = 1 10 CS61C L5 Memory Management; Intro MIPS (16)

Reference Counting Example For every chunk of dynamically allocated memory, keep a count of number of pointers that point to it. When the count reaches 0, reclaim. int *p1, *p2; p1 = malloc(sizeof(int)); p2 = malloc(sizeof(int)); *p1 = 10; *p2 = 20; p1 = p2; Reference count = 2 20 p1 p2 Reference count = 0 10 CS61C L5 Memory Management; Intro MIPS (17)

Reference Counting (p1, p2 are pointers) p1 = p2; Increment reference count for p2 If p1 held a valid value, decrement its reference count If the reference count for p1 is now 0, reclaim the storage it points to. If the storage pointed to by p1 held other pointers, decrement all of their reference counts, and so on Must also decrement reference count when local variables cease to exist. CS61C L5 Memory Management; Intro MIPS (18)

Reference Counting Flaws Extra overhead added to assignments, as well as ending a block of code. Does not work for circular structures! E.g., doubly linked list: X Y Z CS61C L5 Memory Management; Intro MIPS (19)

Scheme 2: Mark and Sweep Garbage Col. Keep allocating new memory until memory is exhausted, then try to find unused memory. Consider objects in heap a graph, chunks of memory (objects) are graph nodes, pointers to memory are graph edges. Edge from A to B => A stores pointer to B Can start with the root set, perform a graph traversal, find all usable memory! 2 Phases: (1) Mark used nodes;(2) Sweep free ones, returning list of free nodes CS61C L5 Memory Management; Intro MIPS (20)

Mark and Sweep Graph traversal is relatively easy to implement recursively void traverse(struct graph_node *node) { /* visit this node */ foreach child in node->children { traverse(child); } } But with recursion, state is stored on the execution stack. Garbage collection is invoked when not much memory left As before, we could traverse in constant space (by reversing pointers) CS61C L5 Memory Management; Intro MIPS (21)

Scheme 3: Copying Garbage Collection Divide memory into two spaces, only one in use at any time. When active space is exhausted, traverse the active space, copying all objects to the other space, then make the new space active and continue. Only reachable objects are copied! Use forwarding pointers to keep consistency Simple solution to avoiding having to have a table of old and new addresses, and to mark objects already copied (see bonus slides) CS61C L5 Memory Management; Intro MIPS (22)

Peer Instruction A. Of {K&R, Slab, Buddy}, there is no best (it depends on the problem). B. Since automatic garbage collection can occur any time, it is more difficult to measure the execution time of a Java program vs. a C program. C. We don t have automatic garbage collection in C because of efficiency. CS61C L5 Memory Management; Intro MIPS (23) ABC 1: FFF 2: FFT 3: FTF 4: FTT 5: TFF 6: TFT 7: TTF 8: TTT

And in semi-conclusion Several techniques for managing heap via malloc and free: best-, first-, next-fit 2 types of memory fragmentation: internal & external; all suffer from some kind of frag. Each technique has strengths and weaknesses, none is definitively best Automatic memory management relieves programmer from managing memory. All require help from language and compiler Reference Count: not for circular structures Mark and Sweep: complicated and slow, works Copying: Divides memory to copy good stuff CS61C L5 Memory Management; Intro MIPS (24)

Forwarding Pointers: 1 st copy abc abc def abc? xyz From To CS61C L5 Memory Management; Intro MIPS (25)

Forwarding Pointers: leave ptr to new abc abc def abc? xyz From To CS61C L5 Memory Management; Intro MIPS (26)

Forwarding Pointers : now copy xyz Forwarding pointer def abc? xyz From To CS61C L5 Memory Management; Intro MIPS (27)

Forwarding Pointers: leave ptr to new xyz Forwarding pointer def abc xyz xyz From To CS61C L5 Memory Management; Intro MIPS (28)

Forwarding Pointers: now copy def Forwarding pointer def abc Forwarding pointer xyz From To Since xyz was already copied, def uses xyz s forwarding pointer to find its new location CS61C L5 Memory Management; Intro MIPS (29)

Forwarding Pointers Forwarding pointer def abc def Forwarding pointer xyz From To Since xyz was already copied, def uses xyz s forwarding pointer to find its new location CS61C L5 Memory Management; Intro MIPS (30)

Assembly Language Basic job of a CPU: execute lots of instructions. Instructions are the primitive operations that the CPU may execute. Different CPUs implement different sets of instructions. The set of instructions a particular CPU implements is an Instruction Set Architecture (ISA). Examples: Intel 80x86 (Pentium 4), IBM/Motorola PowerPC (Macintosh), MIPS, Intel IA64,... CS61C L5 Memory Management; Intro MIPS (31)

Book: Programming From the Ground Up A new book was just released which is based on a new concept - teaching computer science through assembly language (Linux x86 assembly language, to be exact). This book teaches how the machine itself operates, rather than just the language. I've found that the key difference between mediocre and excellent programmers is whether or not they know assembly language. Those that do tend to understand computers themselves at a much deeper level. Although [almost!] unheard of today, this concept isn't really all that new -- there used to not be much choice in years past. Apple computers came with only BASIC and assembly language, and there were books available on assembly language for kids. This is why the old-timers are often viewed as 'wizards': they had to know assembly language programming. -- slashdot.org comment, 2004-02-05 CS61C L5 Memory Management; Intro MIPS (32)

Instruction Set Architectures Early trend was to add more and more instructions to new CPUs to do elaborate operations VAX architecture had an instruction to multiply polynomials! RISC philosophy (Cocke IBM, Patterson, Hennessy, 1980s) Reduced Instruction Set Computing Keep the instruction set small and simple, makes it easier to build fast hardware. Let software do complicated operations by composing simpler ones. CS61C L5 Memory Management; Intro MIPS (33)

MIPS Architecture MIPS semiconductor company that built one of the first commercial RISC architectures We will study the MIPS architecture in some detail in this class (also used in upper division courses CS 152, 162, 164) Why MIPS instead of Intel 80x86? MIPS is simple, elegant. Don t want to get bogged down in gritty details. MIPS widely used in embedded apps, x86 little used in embedded, and more embedded computers than PCs CS61C L5 Memory Management; Intro MIPS (34)

Assembly Variables: Registers (1/4) Unlike HLL like C or Java, assembly cannot use variables Why not? Keep Hardware Simple Assembly Operands are registers limited number of special locations built directly into the hardware operations can only be performed on these! Benefit: Since registers are directly in hardware, they are very fast (faster than 1 billionth of a second) CS61C L5 Memory Management; Intro MIPS (35)

Assembly Variables: Registers (2/4) Drawback: Since registers are in hardware, there are a predetermined number of them Solution: MIPS code must be very carefully put together to efficiently use registers 32 registers in MIPS Why 32? Smaller is faster Each MIPS register is 32 bits wide Groups of 32 bits called a word in MIPS CS61C L5 Memory Management; Intro MIPS (36)

Assembly Variables: Registers (3/4) Registers are numbered from 0 to 31 Each register can be referred to by number or name Number references: $0, $1, $2, $30, $31 CS61C L5 Memory Management; Intro MIPS (37)

Assembly Variables: Registers (4/4) By convention, each register also has a name to make it easier to code For now: $16 - $23 $s0 - $s7 (correspond to C variables) $8 - $15 $t0 - $t7 (correspond to temporary variables) Later will explain other 16 register names In general, use names to make your code more readable CS61C L5 Memory Management; Intro MIPS (38)

C, Java variables vs. registers In C (and most High Level Languages) variables declared first and given a type Example: int fahr, celsius; char a, b, c, d, e; Each variable can ONLY represent a value of the type it was declared as (cannot mix and match int and char variables). In Assembly Language, the registers have no type; operation determines how register contents are treated CS61C L5 Memory Management; Intro MIPS (39)

And in Conclusion In MIPS Assembly Language: Registers replace C variables One Instruction (simple operation) per line Simpler is Better Smaller is Faster New Registers: C Variables: $s0 - $s7 Temporary Variables: $t0 - $t7 CS61C L5 Memory Management; Intro MIPS (40)