SYMBOL TABLE CHAPTER CHAPTER HIGHLIGHTS. 9.1 Operation on Symbol Table. 9.2 Symbol Table Implementation. 9.3 Data Structure for Symbol Table

Similar documents
UNIT-4 (COMPILER DESIGN)

CS1622. Semantic Analysis. The Compiler So Far. Lecture 15 Semantic Analysis. How to build symbol tables How to use them to find

The Structure of a Syntax-Directed Compiler

CA Compiler Construction

The basic operations defined on a symbol table include: free to remove all entries and free the storage of a symbol table

A simple syntax-directed

CS 4240: Compilers and Interpreters Project Phase 1: Scanner and Parser Due Date: October 4 th 2015 (11:59 pm) (via T-square)

SEMANTIC ANALYSIS TYPES AND DECLARATIONS

Semantic Analysis and Type Checking

2.2 Syntax Definition

The compilation process is driven by the syntactic structure of the program as discovered by the parser

Programming Languages Third Edition. Chapter 10 Control II Procedures and Environments

The Structure of a Syntax-Directed Compiler

Single-pass Static Semantic Check for Efficient Translation in YAPL


COMPILER DESIGN. For COMPUTER SCIENCE

Acknowledgement. CS Compiler Design. Semantic Processing. Alternatives for semantic processing. Intro to Semantic Analysis. V.

A Short Summary of Javali

PRINCIPLES OF COMPILER DESIGN UNIT I INTRODUCTION TO COMPILERS

When do We Run a Compiler?

Topic IV. Parameters. Chapter 5 of Programming languages: Concepts & constructs by R. Sethi (2ND EDITION). Addison-Wesley, 1996.

Reading Assignment. Scanner. Read Chapter 3 of Crafting a Compiler.

Symbol Tables. ASU Textbook Chapter 7.6, 6.5 and 6.3. Tsan-sheng Hsu.

Anatomy of a Compiler. Overview of Semantic Analysis. The Compiler So Far. Why a Separate Semantic Analysis?

Topic IV. Block-structured procedural languages Algol and Pascal. References:

Introduction to Compilers and Language Design Copyright (C) 2017 Douglas Thain. All rights reserved.

About the Tutorial. Audience. Prerequisites. Copyright & Disclaimer. Compiler Design

The role of semantic analysis in a compiler

Compiler Theory. (Semantic Analysis and Run-Time Environments)

Chapter 3. Describing Syntax and Semantics

CA Compiler Construction

Compiler construction

Run-Time Data Structures

Semantic Analysis. Outline. The role of semantic analysis in a compiler. Scope. Types. Where we are. The Compiler so far

for (i=1; i<=100000; i++) { x = sqrt (y); // square root function cout << x+i << endl; }

The PCAT Programming Language Reference Manual

Introduction to Lexical Analysis

Semantic Analysis. Outline. The role of semantic analysis in a compiler. Scope. Types. Where we are. The Compiler Front-End

Introduction to Compiler Construction

Syntax Errors; Static Semantics

CS 4201 Compilers 2014/2015 Handout: Lab 1

Chapter 3:: Names, Scopes, and Bindings (cont.)

Introduction to Compiler Construction

Maciej Sobieraj. Lecture 1

Chapter 3:: Names, Scopes, and Bindings (cont.)

for (i=1; i<=100000; i++) { x = sqrt (y); // square root function cout << x+i << endl; }

Defining Program Syntax. Chapter Two Modern Programming Languages, 2nd ed. 1

CSE 307: Principles of Programming Languages

Compiler Design. Subject Code: 6CS63/06IS662. Part A UNIT 1. Chapter Introduction. 1.1 Language Processors

A Simple Syntax-Directed Translator

Introduction to Syntax Analysis. Compiler Design Syntax Analysis s.l. dr. ing. Ciprian-Bogdan Chirila

GBIL: Generic Binary Instrumentation Language. Language Reference Manual. By: Andrew Calvano. COMS W4115 Fall 2015 CVN

Introduction to Compiler Construction

Compilers and Code Optimization EDOARDO FUSELLA

1 Lexical Considerations

Examples of attributes: values of evaluated subtrees, type information, source file coordinates,

Interpreter. Scanner. Parser. Tree Walker. read. request token. send token. send AST I/O. Console

CSE 12 Abstract Syntax Trees

Chapter 3. Describing Syntax and Semantics ISBN

COMPILER CONSTRUCTION LAB 2 THE SYMBOL TABLE. Tutorial 2 LABS. PHASES OF A COMPILER Source Program. Lab 2 Symbol table

CSC488S/CSC2107S - Compilers and Interpreters. CSC 488S/CSC 2107S Lecture Notes

Programming Languages Third Edition. Chapter 7 Basic Semantics

G Programming Languages - Fall 2012

Data Structure for Language Processing. Bhargavi H. Goswami Assistant Professor Sunshine Group of Institutions

Draw a diagram of an empty circular queue and describe it to the reader.

Chapter 3. Describing Syntax and Semantics

Lexical Considerations

5. Semantic Analysis!

5. Semantic Analysis. Mircea Lungu Oscar Nierstrasz

5. Control Statements

Compiler Design Overview. Compiler Design 1

Principles of Programming Languages [PLP-2015] Detailed Syllabus

Instantiation of Template class

The SPL Programming Language Reference Manual

UNIT 3

Intermediate Code Generation

Compilers and Interpreters

COMP-421 Compiler Design. Presented by Dr Ioanna Dionysiou

Objectives. In this chapter, you will:

Parsing and Pattern Recognition

CS 536 Introduction to Programming Languages and Compilers Charles N. Fischer Lecture 11

Compilers. Prerequisites

Introduction to C++ with content from

11. a b c d e. 12. a b c d e. 13. a b c d e. 14. a b c d e. 15. a b c d e

Formal Languages and Compilers Lecture I: Introduction to Compilers

CSE P 501 Compilers. Static Semantics Hal Perkins Winter /22/ Hal Perkins & UW CSE I-1

Chapter 12 Variables and Operators

Compiler Design. Computer Science & Information Technology (CS) Rank under AIR 100

EI326 ENGINEERING PRACTICE & TECHNICAL INNOVATION (III-G) Kenny Q. Zhu Dept. of Computer Science Shanghai Jiao Tong University

CSE 504. Expression evaluation. Expression Evaluation, Runtime Environments. One possible semantics: Problem:

LECTURE 18. Control Flow

COSC160: Data Structures Hashing Structures. Jeremy Bolton, PhD Assistant Teaching Professor

CSC 467 Lecture 13-14: Semantic Analysis

Understand how to deal with collisions

CS /534 Compiler Construction University of Massachusetts Lowell. NOTHING: A Language for Practice Implementation

Yacc Meets C+ + Stephen C. Johnson Ardent Computer Corporation

Introduction. hashing performs basic operations, such as insertion, better than other ADTs we ve seen so far

COMPILERS BASIC COMPILER FUNCTIONS

CS164: Programming Assignment 2 Dlex Lexer Generator and Decaf Lexer

Short Notes of CS201

Transcription:

CHAPTER 9 SYMBOL TABLE CHAPTER HIGHLIGHTS 9.1 Operation on Symbol Table 9.2 Symbol Table Implementation 9.3 Data Structure for Symbol Table 9.3.1 List 9.3.2 Self Organizing List 9.3.3 Hash Table 9.3.4 Binary Search Tree 9.4 Symbol Table Handler Tribulations Further Readings and References

292 Design and Implementation of Compiler After syntax tree have been constructed, the compiler must check whether the input program is typecorrect (called type checking and part of the semantic analysis). During type checking, a compiler checks whether the use of names (such as variables, functions, type names) is consistent with their definition in the program. Consequently, it is necessary to remember declarations so that we can detect inconsistencies and misuses during type checking. This is the task of a symbol table. Note that a symbol table is a compile-time data structure. It s not used during run time by statically typed languages. Formally, a symbol table maps names into declarations (called attributes), such as mapping the variable name x to its type int. More specifically, a symbol table stores: For each type name, its type definition. For each variable name, its type. If the variable is an array, it also stores dimension information. It may also store storage class, offset in activation record etc. For each constant name, its type and value. For each function and procedure, its formal parameter list and its output type. Each formal parameter must have name, type, type of passing (by-reference or by-value), etc. 9.1 OPERATION ON SYMBOL TABLE We need to implement the following operations for a symbol table: 1. insert ( String key, Object binding ) 2. object_lookup ( String key ) 3. begin_scope () and end_scope () (1) insert (s,t)- return index of new entry for string s and token t (2) lookup (s)- return index of new entry for string s or o if s is not found. (3) begin_scope () and end_scope () : When we have a new block (ie, when we encounter the token {), we begin a new scope. When we exit a block (i.e. when we encounter the token }) we remove the scope (this is the end_scope). When we remove a scope, we remove all declarations inside this scope. So basically, scopes behave like stacks. One way to implement these functions is to use a stack. When we begin a new scope we push a special marker to the stack (e.g., 1). When we insert a new declaration in the hash table using insert, we also push the bucket number to the stack. When we end a scope, we pop the stack until and including the first 1 marker. Example 9.1: Consider the following program: 1) { 2) int a; 3) { 4) int a; 5) a = 1; 6) }; 7) a = 2; 8) }; we have the following sequence of commands for each line in the source program (we assume that the hash key for a is 12): 1) push( 1) 2) insert the binding from a to int into the beginning of the list table[12] push(12) 3) push( 1) 4) insert the binding from a to int into the beginning of the list table[12] Contd...

Symbol Table 293 push(12) 6) pop() remove the head of table[12] pop() 7) pop() remove the head of table[12] pop() Recall that when we search for a declaration using lookup, we search the bucket list from the beginning to the end, so that if we have multiple declarations with the same name, the declaration in the innermost scope overrides the declaration in the outer scope. (4) Handling Reserve Keywords: Symbol table also handle reserve keywords like PLUS, MINUS, MUL etc. This can be done in following manner. insert ( PLUS, PLUS); insert ( MINUS, MINUS); In this case first PLUS and MINUS indicate lexeme and other one indicate token. 9.2 SYMBOL TABLE IMPLEMENTATION The data structure for a particular implementation of a symbol table is sketched in figure 9.1. In figure 9.1, a separate array arr_lexemes holds the character string forming an identifier. The string is terminated by an end-of-string character, denoted by EOS, that may not appear in identifiers. Each entry in symbol-table array arr_symbol_table is a record consisting of two fields, as lexeme_pointer, pointing to the beginning of a lexeme, and token. Additional fields can hold attribute values. In figure 9.1, the 0th entry is left empty, because lookup return 0 to indicate that there is no entry for a string. The 1 st, 2 nd, 3 rd, 4 th, 5 th, 6 th, and 7 th entries are for the a, plus b and, c, minus, and d where 2 nd, 4 th and 6 th entries are for reserve keyword. ARRAY all_symbol_table a+b AND c d Lexeme_pointer Token Attribute Position 0 id 1 plus 2 id 3 AND 4 id 5 minus 6 id 7 a EOS P L U S EOS b EOS AND EOS c EOS M I N U S EOS d ARRAY arr_lexeme Figure 9.1: Implemented symbol table

294 Design and Implementation of Compiler When lexical analyzer reads a letter, it starts saving letters, digits in a buffer lex_bufffer. The string collected in lex_bufffer is then looked in the symbol table, using the lookup operation. Since the symbol table initialized with entries for the keywords plus, minus, AND operator and some identifiers as shown in figure 9.1 the lookup operation will find these entries if lex_buffer contains either div or mod. If there is no entry for the string in lex_buffer, i.e., lookup return 0, then lex_buffer contains a lexeme for a new identifier. An entry for the identifier is created using insert( ). After the insertion is made; n is the index of the symbol-table entry for the string in lex_buffer. This index is communicated to the parser by setting tokenval to n, and the token in the token field of the entry is returned. 9.3 DATA STRUCTURE FOR SYMBOL TABLE 9.3.1 List The simplest and easiest to implement data structure for symbol table is a linear list of records. We use single array or collection of several arrays for this purpose to store name and their associated information. Now names are added to end of array. End of array always marks by a point known as space. When we insert any name in this list then searching is done in whole array from space to beginning of array. If word is not found in array then we create an entry at space and increment space by one or value of data type. At this time insert ( ), object look up ( ) operation are performed as major operation while begin_scope ( ) and end_scope( ) are used in simple table as minor operation field as token type attribute etc. In implementation of symbol table first field always empty because when object-lookup work then it will return 0 to indicate no string in symbol table. Complexity: If any symbol table has n names then for inserting any new name we must search list n times in worst case. So cost of searching is O(n) and if we want to insert n name then cost of this insert is O(n^2) in worst case. Variable Information(type Space (byte) a b Float 2 4 c d Character Long 1 4 SPACE Figure 9.2: Symbol table as list

Symbol Table 295 9.3.2 Self Organizing List To reduce the time of searching we can add an addition field linker to each record field or each array index. When a name is inserted then it will insert at space and manage all linkers to other existing name. Variable Id1 Information Variable Id1 Information Info1 Info2 Info3 Info1 Info2 Id2 SPACE Id2 Id3 SPACE (a) (b) Figure 9.3: Symbol table as self organizing list In above figure (a) represent the simple list and (b) represent self organzing list in which Id1 is related to Id2 and Id3 is related to Id1. 9.3.3 Hash Table A hash table, or a hash map, is a data structure that associates keys with values Open hashing is a key that is applied to hash table. In hashing open, there is a property that no limit on number of entries that can be made in table. Hash table consist an array HESH and several buckets attached to array HESH according to hash function. Main advantage of hash table is that we can insert or delete any number or name in O (n) time if data are search linearly and there are n memory location where data is stored. Using hash function any name can be search in O(1) time. However, the rare worst-case lookup time can be as bad as O(n). A good hash function is essential for good hash table performance. A poor choice of a hash function is likely to lead to clustering, in which probability of keys mapping to the same hash bucket (i.e. a collision) occur. One organization of a hash table that resolves conflicts is chaining. The idea is that we have list elements of type: class Symbol { public String key; public Object binding; public Symbol next; public Symbol ( String k, Object v, Symbol r ) { key=k; binding=v; next=r; } }

296 Design and Implementation of Compiler Structure of hash table look like as a \n b c d \n e \n f g \n Figure 9.4: Symbol table as hash table (\n represent NULL) 9.3.4 Search Tree Another approach to organize symbol table is that we add two link fields i.e. left and right child, we use these field as binary search tree. All names are created as child of root node that always follow the property of binary tree i.e. name <name ie and Namej <name. These two statements show that all smaller name than Namei must be left child of name otherwise right child of namej. For inserting any name it always follow binary search tree insert algorithm. Example 9.2: Create list, search tree and hash table for given program for given program int a,b,c; int sum (int x, int y) { a = x+y return (a) } main () { int u, u=sum (5,6); }

Symbol Table 297 (i) List Variable Information Space(byte) u a b c x y sum Space Figure 9.5: Symbol table as list for example 9.2 (ii) Hash Table a b c \n sum \n u \n x y \n Figure 9.6: Symbol table as hash table for example 9.2

298 Design and Implementation of Compiler (iii) Search Tree u a x b y c sum Figure 9.7: Symbol table as search tree for example 9.2 9.4 SYMBOL TABLE HANDLER Any interface i.e. symbol table between source and object code must take recognisance of data-related concepts like storage, addresses and data representation as well as control-related ones like location counter, sequential execution and branch instruction which are fundamental to nearly all machines on which highlevel language programs execute. Typically machines allow some operations which simulate arithmetic or logical operations on data bit patterns which simulate numbers or characters, these patterns being stored in an array like structure of memory whose elements are distinguished by addresses. In high-level languages these addresses are usually given mnemonic names. The context-free syntax of many high-level languages seems to draw a distinction between the address for a variable and the value associated with that variable and stored at its address. Hence we find statements like X := X + 4 in which the X on the left of the := operator actually represents an address (sometimes called the Lvalue of X ) while the X on the right (sometimes called the R-value of X ) actually represents the value of the quantity currently residing at the same address or we can say that each X i.e., all variable in the above assignment was syntactically a Designator. Semantically these two designators are very different we shall refer to the one that represents an address as a Variable Designator and to the one that represents a value as a Value Designator. To perform its task, the code generation interface will require the extraction of further information associated with user-defined identifiers and best kept in the symbol table. In the case of constants we need to record the associated values and in the case of variables we need to record the associated addresses and storage demands (the elements of array variables will occupy a contiguous block of memory).

Symbol Table 299 Handling the different manners of entries that need to be stored in a symbol table can be done in various ways (described in section 9.4). One different method from 9.4 is object-oriented class based implementation one might define an abstract base class to represent a generic type of entry and then derive classes from this to represent entries for variables or constants. The traditional way, still required if one is hosting a compiler in a language that does not support inheritance as a concept, is to make use of union (in C++ terminology). Since the class-based implementation gives so much scope for exercises, we have chosen to illustrate the variant record approach which is very efficient and quite adequate for such a simple language. We extend the declaration of the TABLE_entries type to be struct TABLE_entries { TABLE_alfa name; // identifier TABLE_idclasses idclass; // class union { struct { int value; } c; // constants struct { int size, offset; // number of words, relative address bool scalar; // distinguish arrays } v; // variables }; }; TRIBULATIONS 9.1 How would you check that no identifier is declared more than once? 9.2 How do real compilers deal with symbol tables? 9.3 How do real compilers keep track of type checking via symbol table? Why should name equivalence be easier to handle than structural equivalence? 9.4 Why do some languages simply prohibit the use of anonymous types and why don t more languages forbid them? 9.5 How do you suppose compilers keep track of storage allocation for struct or RECORD types, and for union or variant record types? 9.6 Find out how storage is managed for dynamically allocated variables in language like C++. 9.7 How does one cope with arrays of variable (dynamic) length in subprograms? 9.8 Identifiers that are undeclared by virtue of mistyped declarations tend to be trying for they result in many subsequent errors being reported. Perhaps in languages as simple as ours one could assume that all undeclared identifiers should be treated as variables and entered as such in the symbol table at the point of first reference. Is this a good idea? Can it easily be implemented? What happens if arrays are undeclared? 9.9 C language array declaration is different from a C++ one the bracketed number in C language specifies the highest permitted index value, rather than the array length. This has been done so that one can declare variables like VAR Scalar, List[10], VeryShortList[0];

300 Design and Implementation of Compiler How would you modify C language and Topsy to use C++ semantics, where the declaration of VeryShortList would have to be forbidden? 9.10 Yet another approach is to construct the symbol table using a hash table which probably yields the shortest times for retrievals. Develop a hash table implementation for your C compiler. 9.11 We might consider letting the scanner interact with the symbol table. Develop a scanner that stores the strings for identifiers and string literals in a string table. 9.12 Develop a symbol table handler that utilizes a simple class hierarchy for the possible types of entries inheriting appropriately from a suitable base class. Once again construction of such a table should prove to be straightforward regardless of whether you use a linear array, tree or hash table as the underlying storage structure. Retrieval might call for more initiative since C++ does not provide syntactic support for determining the exact class of an object that has been statically declared to be of a base class type. 9.13 Why can we easily allow the declaration of a pointer type to precede the definition of the type it points to even in a one-pass system? 9.14 One might accuse the designers of Pascal and C of making a serious error of judgement they do not introduce a string type as standard but rely on programmers to manipulate arrays of characters and to use error level ways of recognizing the end or the length of a string. Do you agree? Discuss whether what they offer in return is adequate and if not why not. Suggest why they might deliberately not have introduced a string type. 9.15 Many authors dislike pointer types because they allow insecure programming. What is meant by this? How could the security be improved? If you do not like pointer types, can you think of any alternative feature that would be more secure? FURTHER READINGS AND REFERENCES [1] Alfred V. Aho, Ravi Sethi, and Jeffrey D. Ullman. Compilers: Principles, Techniques and Tools. [2] Holub, Allen. Compiler Design in C Extensive examples in C. [3] Appel, Andrew. Modern Compiler Implementation in C/Java/ML is a set of cleanly written texts on compiler design, studied from various different methodological perspectives. [4] Brown, P.J. Writing Interactive Compilers and Interpreters. Useful practical advice, not much theory. [5] Fischer, Charles & LeBlanc, Richard. Crafting A Compiler. Uses an ADA like pseudo-code. [6] Weinberg, G.M. The Psychology of Computer Programming: Silver Anniversary. Interesting insights and anecdotes. [7] Wirth, Niklaus. Compiler Construction. From the inventor of Pascal, Modula-2 and Oberon-2, examples in Oberon. [8] Compiler textbook references A collection of references to mainstream Compiler Construction Textbook. [9] Wirth, Niklaus Compiler Construction, Addison-Wesley 1996, pp.176. [10] Cifuentes, C. Reverse Compilation Techniques. PhD thesis, Queensland University of Technology, 1994. (available as compressed postscript). [11] Czarnota, B. and Hart R.J. Legal protection of computer programs in Europe: a guide to the EC directive. 1991, London: Butterworths. [12] Terry, P.D. Compilers and Compiler Generators an introduction with C++, Rhodes University, 1996.