Ling/CMSC 773 Take-Home Midterm Spring 2007

Size: px
Start display at page:

Download "Ling/CMSC 773 Take-Home Midterm Spring 2007"

Transcription

1 Ling/CMSC 773 Take-Home Midterm Spring 2007 Ground rules. This is a take-home exam. Each person should be working completely independently no communication of any kind. There will be no communication with the instructor or TA either. If you have questions or believe that there is an error of some kind, do your best to answer, making whatever assumptions you feel are appropriate and necessary and saying explicitly what they are. Exams must be turned in either by to Philip, or by hardcopy under Philip s office door, AVW3143; in both cases the deadline for receipt is Monday, April 2, at 9am. (Philip or an assistant will be picking up the hardcopies at 9am sharp.) Use of notes and other materials. This exam is open book: you are free to use the textbook, your notes, and any other hardcopy materials (even, say, another textbook). However, you may not use the Web. Once the exam is handed out, and until it is turned in, you are forbidden to search the Web for explanations, notes, tutorials, or any other materials. The only Web pages you are permitted to access in relation to this examination are which contains errata for the textbook, and any page explicitly linked from the course Web page or syllabus. (But not the transitive closure of that linking relationship!) Programming. This exam is to be done entirely without computer programming. The only permitted use of software is for word processing (e.g. MS Word, LaTeX, etc.) or related creation of charts/graphs. Use of spreadsheet software (e.g. Excel) to do numerical calculations is not permitted, although you can use Excel to create graphs and charts if so desired. (The use of calculators is, of course, allowed.) Honor pledge. On the first page of your examination, in addition to your name, you are expected to include and sign (or type, for e-submissions) the following: I pledge on my honor that I have not given or received any unauthorized assistance on this examination. For purposes of this exam unauthorized assistance includes access to non-permitted notes, computer programs, or other materials as specified above. Other notes. Plesae turn in your answers in the same order that questions appear on the exam. Make sure to clearly identify which answers go with which questions, even for sub-parts of questions. (I don t want to have to guess where you make the transition from part (a) to part (b), etc.) You are encouraged to show all your work, so that you can be awarded partial credit. You are welcome to turn in material that includes scratch paper or scratch space, if you like, but if you do so then you need to clearly indicate what is intended to be graded. Unless the question says otherwise, for numerical calculations, you do not need to do the arithmetic for example, in most cases, if you re computing a product of probabilities, you can give your answer in the form , or whatever. Just make sure it s completely clear where the numbers come from. Please return this document to the instructor at the next class, with your name clearly written at the top. 1

2 1 Short Answer Questions [30 points total] Answer each question clearly and briefly. (Even if you make sure to be pretty thorough which you should, in order to obtain partial credit these are short answers, not long essays.) (a) Give a constituency tree for Many students said they enjoyed the interesting exam (using the usual sort of notation from class, please, nothing linguistically fancy). Give a head table for the context-free rules needed to construct the tree, i.e. context-free rules annotated to show which constituent is the head. Then show explicitly, step-by-step, how to convert your constituency tree into an unlabeled dependency tree. (b) For this question, assume data sparseness is not a problem, i.e. assume all probabilities can be estimated accurately. Consider the difficulty that a simple probabilistic CFG has in distinguishing the likelihood of the sentences people eat roasted peanuts and peanuts eat roasted people. Briefly describe one solution to this problem and its advantages and disadvantages. Now consider the sentence Children sometimes munch chocolate soldiers on Christmas morning. With the solution you just described, is the probability of this sentence likely to have a value that accords with your intuition? Explain why or why not. (c) Suppose you have an unfair coin that comes up heads 7/8 of the time when you flip it. Suppose you flip the coin 1000 times and report the outcome each time, paying one penny per bit for your communication. If you re thrifty (you want to pay as little as possible) and smart (you ve taken this class), what s the lowest amount you could expect to get away with paying? Justify your answer. (d) A former Amazon employee once said: We sold more books today that didn t sell at all yesterday than we sold today of all the books that did sell yesterday. (To state this in plainer English, If you look at our infrequently-sold books, the total number we sell (let s call it N) is very large, even if each individual book is only bought by a few people. And if you look at N, it s bigger than the total sales of the frequently sold books! ) George Zipf would be unsurprised. Briefly explain why. (e) Briefly explain/illustrate how a measure of uncertainty can also be considered a measure of surprise and a measure of quantity of information. (f) Here s an argument: Consider the sentences (i) Hold the newsreader s nose squarely, waiter, or friendly milk will countermand my trousers, and (ii) Countermand friendly, hold milk my, newsreader s nose or squarely the trousers waiter will. Neither (it s fair to assume) has ever occurred in the prior experience of a speaker of English, and yet any such speaker would readily identify the former as grammatical and the latter as ungrammatical. Therefore one cannot associate the notion grammaticality of a sentence in English with the notion likelihood of a sentence in English. Briefly but convincingly demolish this argument. 2

3 2 EM [20 points] Consider the following hidden model variation on a bigram model for word sequences. As in the usual bigram model, we express the probability of an entire sequence w 1 w 2... w T by (1) T p(w 1 w 2... w T ) = p(w 1 ) p(w t w t 1 ). t=2 However, the parameters used in the product are defined as follows: (2) p(w t w t 1 ) = C p(w t c)p(c w t 1 ) i=1 In plain English, the generative story for this model is the following. Instead of generating the next word w t based on the previous word w t 1, as in a usual bigram model, we generate a class c based on w t 1, and then we generate w t based on c. 1 So the probability of choosing the next word w t is a sum of probabilities, one for each hidden class. The probability contributed to the choice of w t for each class c, namely p(w t c)p(c w t 1 ), represents the joint probability of two events: picking c based on w t 1, and then picking w t based on c. Since c is hidden, we have to sum up over the different possibilities. Intuitively, the hidden class c can be viewed as capturing the general properties of w t 1 that are relevant for generating the next word. Or you can think of going through hidden class c to get from w t 1 to w t. As you can see, there are two sets of parameters in this model. The first set is the word-to-class probabilities p(c w i ), and the second set is the class-to-word probabilities p(w j c). (Where both w i and w j range over the entire vocabulary, and c ranges over the set of C hidden classes.) Because of the way the model is structured, there s an EM algorithm for estimating these parameters that is much simpler than the Forward-Backward algorithm. In particular, there s no need at all for dynamic programming. 2 For the sake of consistent notation, please use N(x, y) as your notation for counts. E.g. N(w i, w j ) would be the number of times w i is followed by w j. (a) (2 points) In addition to capturing generalizations (by associating words with abstract classes that are learned automatically), an advantage of this model is that in practice it has far fewer parameters than a usual (non-hidden) bigram model. Let the size of the vocabulary be V and the number of classes be C. How many classes can you use in the hidden model before the total number of parameters exceeds the number of parameters in the usual (non-hidden) bigram model? If you re comparing the total number of parameters in the two models, What does the comparison look like for typical values of C = 32 and V = 50000? (b) (2 points) Suppose the classes c were observable rather than hidden. Express the maximum likelihood estimate for the probability p(w j c) in terms of counts. (c) (2 points) Suppose the classes c were observable rather than hidden. Express the maximum likelihood estimate for the probability p(c w i ) in terms of counts. (d) (7 points) Recall that for many cases of EM algorithms, the basic structure of the algorithm can be described as follows: 1. Set initial values for parameters µ 1 Also, as usual, we can assume w 1 is always a special start word that always starts an observed sequence with probability 1. 2 Intuitively, this is because the choice of the hidden class at every step depends only on what s observable, i.e. hidden classes are independent of each other given the intervening word. 3

4 2. E-step: Figure out expected counts of relevant events, where those events typically involve both observable and hidden values, using the current parameters µ to determine what s expected. 3. M-step: Use the expected counts to compute µ new, a new set of parameters. In the EM algorithms we ve studied, this is a maximum likelihood estimate, i.e. just normalizing (expected) counts. 4. If the probabilities have converged (or if we ve done some maximum number of iterations), stop; otherwise let µ = µ new and go back to the E-step for the next iteration. If you read about the EM algorithm for this model, the updating of the parameters is described as having the following M-step for the two sets of parameters. w p new (c w i ) = j N(w i, w j )p(c w i, w j ) (3) c w j N(w i, w j )p(c w i, w j ) w p new (w j c) = i N(w i, w j )p(c w i, w j ) (4) w i N(w i, w j )p(c w i, w j ) The E-step is described as: (5) w j p(c w i, w j ) = p(w j c)p(c w i ) c p(w j c )p(c w i ) In words, p(c w i, w j ) can be thought of as the probability that c was the hidden state that got used when going from w i to w j. Explain why equation 4 is the way to calculate the new value for p new (w j c) given the previous guesses for the parameters. Partial credit will be assigned for good explanations, but a perfect answer will show via equations (along with written explanation) how to derive equation 4 using the non-hidden maximum likelihood estimate (part b) as the starting point. (I.e. the form of the explanation will follow the same general schema we used for deriving the update for HMM transition probabilities a ij, starting with the maximum likelihood estimate you d use if all the state transitions were visible. Though, as was mentioned, no dynamic program is needed.) (e) (7 points) Same as part (d), but explain why equation 3 is the way to update p new (c w i ). For those who have ever looked more formally at EM: you will not receive credit for deriving the EM updates from the model s likelihood function, using Lagrange multipliers to optimize, etc. I don t want to see a Q function. I m looking for an understanding of the updates in relation to the non-hidden maximum likelihood estimate, as was our focus in discussions of the Forward-Backward and Inside-Outside algorithms. 4

5 3 N-gram models [15 points] Consider the following problem, known as language identificiation : given a previously unseen string of natural language text, what language is it written in? This is often solved using n-gram models. Assume you have training samples of 20,000 words in each of k languages, {L 1,..., L k }. Furthermore, to keep things simple, assume that all of these languages are alphabetic (i.e. not ideographic like many Asian languages) and that all data is in Unicode, and that all k languages use significantly overlapping alphabets (e.g. you could suppose they all use a roman alphabet with accents/diacritics).. If you need to for concreteness, you can assume the genre is newswire, and that there is no markup. Now, imagine that you are asked to design a solution to the language identification problem based on n-grams. (a) (5 points) Would you use n-grams composed of words or characters? Justify your answer in quantitative terms, clearly stating any assumptions you make (e.g. about sizes of vocabularies, sizes of alphabets, or any other relevant quantities). (b) (5 points) Explain in detail how, given a new piece of text in one of the k languages, you would take a Bayesian approach to identifying which language L i it is written in. (c) (5 points) Suppose you have read about two different kinds of n-gram model (e.g. a bigram model and a trigram model, or a standard bigram model and the aggregate bigram model of the previous question) and you want to know which kind performs better on this problem. Under the assumption that you have only the small amount of data in each language as described above, explain in detail how you would conduct an evaluation to assess which of the two kinds of models is better for this problem. 5

6 4 Relative Entropy [20 points] Selectional preferences are semantic constraints that a predicate places on its arguments. For example, the verb drink prefers that its objects be in the semantic class beverage, which is why John drank the wine sounds a lot better than, say, John drank the toaster. In the early 1990s, the following was proposed as a probabilistic model of selectional preferences: S object (v) = D(p(c v) p(c)) = Pr(object is in class c verb is v) log c Classes A object (v, c) = Pr(object is in class c verb is v) log Pr(object is in class c verb is v). Pr(object is in class c) Pr(object is in class c verb is v). Pr(object is in class c) The value S object (v) was referred to as the selectional preference strength of verb v, and the value A object (v, c) was called the selectional association between the verb v and a particular semantic category c. (a) Assuming probabilities are estimated accurately (for example, pretend the classes are observable), and assuming that verb-object frequencies in the corpus are a reasonably accurate reflection of the real world, would the model correctly predict that A object (drink, beverage) should be higher than A object (drink, appliance)? Explain why or why not. (b) Below, the graph on the left shows A object (v 1, c) and the graph on the right shows A object (v 2, c), for two verbs v 1 and v 2. (The x-axis ranges over possible semantic categories c, and each bar gives the value of A object (v, c).) As you can see, one of the verbs is eat and the other one is find. Even without the labels, you would have been able to identify which was which, based on what you know about the model (and the assumptions we ve made). Explain how, with explicit reference to the formal definition of the model. A(v,c) x Selectional Profile: find profile.find Class ID x 10 6 A(v,c) x Selectional Profile: eat profile.eat Class ID x (c) The person who proposed this model claimed that S object (v) models, quantitively, how much information the verb carries about the semantic category of its direct object. Explain why this claim is true. 6

7 5 Parsing and Evaluation [15 points] You have been hired by a AwesomeSearchEngine.com to make recommendations concerning language technology. They have been approached by Whizdee, Inc., a startup company, and in a sales presentation Whizdee s sales representative says the following: We re doing very exciting work on parsing, and our results are very impressive. In one experiment, we trained one of our parsers on 80% of the Penn Treebank and tested on the other 20%, and its labeled recall on constituent boundaries was 98.2%. We also did an experiment, training on the same data, where the labeled precision for constituent boundaries was 97.6%. With numbers like that, how could you lose? (a) Consider the true parse T, and the system parse P, below: [T] (S (NP the prof) (VP (VPRT looked up) (NP the grade)) (ADVP today)) [P] (S (NP the prof) (VP looked (PP up (NP the grade)) (ADVP today))) What are the values for labeled precision and labeled recall? Note that I ve omitted all part of speech labels because they re not used in constituent recall/precision calculations. Every uppercase symbol in the parse is a constituent label. (b) Should people be impressed by Whizdee s numbers? Explain why or why not. (c) Whizdee s decided you re so smart that they, too, want to pay you as a consultant. They ve got a questionanswering search engine that uses a parser, and a contract from the New York Times. They paid minimum wage to impoverished linguistics grad students in New York City to create parse trees for 20,000 New York Times articles written between March 15, 1997 and March 15, Whizdee plans to make their system available to the public starting April 1, but the New York Times insists on a formal parser evaluation first. One of Whizdee s scientists says that they should evaluate their parser by doing 10-fold cross validation. Another of their scientists says that they should evaluate it by training on the data up to March 15, 2006 and testing on the rest. Explain what the two competing evaluation approaches are and discuss the advantages and disadvantages of each approach. 7

Exam Marco Kuhlmann. This exam consists of three parts:

Exam Marco Kuhlmann. This exam consists of three parts: TDDE09, 729A27 Natural Language Processing (2017) Exam 2017-03-13 Marco Kuhlmann This exam consists of three parts: 1. Part A consists of 5 items, each worth 3 points. These items test your understanding

More information

CS 4349 Lecture August 21st, 2017

CS 4349 Lecture August 21st, 2017 CS 4349 Lecture August 21st, 2017 Main topics for #lecture include #administrivia, #algorithms, #asymptotic_notation. Welcome and Administrivia Hi, I m Kyle! Welcome to CS 4349. This a class about algorithms.

More information

UNIVERSITY OF EDINBURGH COLLEGE OF SCIENCE AND ENGINEERING SCHOOL OF INFORMATICS INFR08008 INFORMATICS 2A: PROCESSING FORMAL AND NATURAL LANGUAGES

UNIVERSITY OF EDINBURGH COLLEGE OF SCIENCE AND ENGINEERING SCHOOL OF INFORMATICS INFR08008 INFORMATICS 2A: PROCESSING FORMAL AND NATURAL LANGUAGES UNIVERSITY OF EDINBURGH COLLEGE OF SCIENCE AND ENGINEERING SCHOOL OF INFORMATICS INFR08008 INFORMATICS 2A: PROCESSING FORMAL AND NATURAL LANGUAGES Saturday 10 th December 2016 09:30 to 11:30 INSTRUCTIONS

More information

CSE 131 Introduction to Computer Science Fall 2016 Exam I. Print clearly the following information:

CSE 131 Introduction to Computer Science Fall 2016 Exam I. Print clearly the following information: CSE 131 Introduction to Computer Science Fall 2016 Given: 29 September 2016 Exam I Due: End of Exam Session This exam is closed-book, closed-notes, no electronic devices allowed The exception is the "sage

More information

Semantics via Syntax. f (4) = if define f (x) =2 x + 55.

Semantics via Syntax. f (4) = if define f (x) =2 x + 55. 1 Semantics via Syntax The specification of a programming language starts with its syntax. As every programmer knows, the syntax of a language comes in the shape of a variant of a BNF (Backus-Naur Form)

More information

Department of Electrical Engineering and Computer Sciences Spring 2001 Instructor: Dan Garcia CS 3 Midterm #2. Personal Information

Department of Electrical Engineering and Computer Sciences Spring 2001 Instructor: Dan Garcia CS 3 Midterm #2. Personal Information University of California, Berkeley College of Engineering Department of Electrical Engineering and Computer Sciences Spring 2001 Instructor: Dan Garcia 2001-04-09 CS 3 Midterm #2 Personal Information Last

More information

Design First ITS Instructor Tool

Design First ITS Instructor Tool Design First ITS Instructor Tool The Instructor Tool allows instructors to enter problems into Design First ITS through a process that creates a solution for a textual problem description and allows for

More information

The Expectation Maximization (EM) Algorithm

The Expectation Maximization (EM) Algorithm The Expectation Maximization (EM) Algorithm continued! 600.465 - Intro to NLP - J. Eisner 1 General Idea Start by devising a noisy channel Any model that predicts the corpus observations via some hidden

More information

Recognizing a Function

Recognizing a Function Recognizing a Function LAUNCH (7 MIN) Before Why would someone hire a dog walking service? During Do you know exactly what it would cost to hire Friendly Dog Walking? After How does each service encourage

More information

Math 7 Notes - Unit 4 Pattern & Functions

Math 7 Notes - Unit 4 Pattern & Functions Math 7 Notes - Unit 4 Pattern & Functions Syllabus Objective: (3.2) The student will create tables, charts, and graphs to extend a pattern in order to describe a linear rule, including integer values.

More information

CS 161 Computer Security

CS 161 Computer Security Popa & Wagner Spring 2016 CS 161 Computer Security Midterm 2 Print your name:, (last) (first) I am aware of the Berkeley Campus Code of Student Conduct and acknowledge that academic misconduct will be

More information

Assignment 4 CSE 517: Natural Language Processing

Assignment 4 CSE 517: Natural Language Processing Assignment 4 CSE 517: Natural Language Processing University of Washington Winter 2016 Due: March 2, 2016, 1:30 pm 1 HMMs and PCFGs Here s the definition of a PCFG given in class on 2/17: A finite set

More information

University of Virginia Department of Computer Science. CS 4501: Information Retrieval Fall 2015

University of Virginia Department of Computer Science. CS 4501: Information Retrieval Fall 2015 University of Virginia Department of Computer Science CS 4501: Information Retrieval Fall 2015 2:00pm-3:30pm, Tuesday, December 15th Name: ComputingID: This is a closed book and closed notes exam. No electronic

More information

University of Toronto Department of Electrical and Computer Engineering. Midterm Examination. ECE 345 Algorithms and Data Structures Fall 2012

University of Toronto Department of Electrical and Computer Engineering. Midterm Examination. ECE 345 Algorithms and Data Structures Fall 2012 1 University of Toronto Department of Electrical and Computer Engineering Midterm Examination ECE 345 Algorithms and Data Structures Fall 2012 Print your name and ID number neatly in the space provided

More information

The CKY Parsing Algorithm and PCFGs. COMP-550 Oct 12, 2017

The CKY Parsing Algorithm and PCFGs. COMP-550 Oct 12, 2017 The CKY Parsing Algorithm and PCFGs COMP-550 Oct 12, 2017 Announcements I m out of town next week: Tuesday lecture: Lexical semantics, by TA Jad Kabbara Thursday lecture: Guest lecture by Prof. Timothy

More information

CSE 494: Information Retrieval, Mining and Integration on the Internet

CSE 494: Information Retrieval, Mining and Integration on the Internet CSE 494: Information Retrieval, Mining and Integration on the Internet Midterm. 18 th Oct 2011 (Instructor: Subbarao Kambhampati) In-class Duration: Duration of the class 1hr 15min (75min) Total points:

More information

Database Management System Prof. Partha Pratim Das Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur

Database Management System Prof. Partha Pratim Das Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur Database Management System Prof. Partha Pratim Das Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur Lecture - 19 Relational Database Design (Contd.) Welcome to module

More information

A Neuro Probabilistic Language Model Bengio et. al. 2003

A Neuro Probabilistic Language Model Bengio et. al. 2003 A Neuro Probabilistic Language Model Bengio et. al. 2003 Class Discussion Notes Scribe: Olivia Winn February 1, 2016 Opening thoughts (or why this paper is interesting): Word embeddings currently have

More information

University of Toronto Department of Computer Science

University of Toronto Department of Computer Science University of Toronto Department of Computer Science CSC340S - Information Systems Analysis and Design March 14, 2003 John Mylopoulos Assignment 3: Detailed Design for an Information System Due Date: 12:00midnight

More information

Computer Science 161

Computer Science 161 Computer Science 161 150 minutes/150 points Fill in your name, logname, and TF s name below. Name Logname The points allocated to each problem correspond to how much time we think the problem should take.

More information

Topics in Parsing: Context and Markovization; Dependency Parsing. COMP-599 Oct 17, 2016

Topics in Parsing: Context and Markovization; Dependency Parsing. COMP-599 Oct 17, 2016 Topics in Parsing: Context and Markovization; Dependency Parsing COMP-599 Oct 17, 2016 Outline Review Incorporating context Markovization Learning the context Dependency parsing Eisner s algorithm 2 Review

More information

Proofwriting Checklist

Proofwriting Checklist CS103 Winter 2019 Proofwriting Checklist Cynthia Lee Keith Schwarz Over the years, we ve found many common proofwriting errors that can easily be spotted once you know how to look for them. In this handout,

More information

6.001 Notes: Section 6.1

6.001 Notes: Section 6.1 6.001 Notes: Section 6.1 Slide 6.1.1 When we first starting talking about Scheme expressions, you may recall we said that (almost) every Scheme expression had three components, a syntax (legal ways of

More information

An Interesting Way to Combine Numbers

An Interesting Way to Combine Numbers An Interesting Way to Combine Numbers Joshua Zucker and Tom Davis October 12, 2016 Abstract This exercise can be used for middle school students and older. The original problem seems almost impossibly

More information

Midterm Exam Fundamentals of Computer Graphics (COMP 557) Thurs. Feb. 19, 2015 Professor Michael Langer

Midterm Exam Fundamentals of Computer Graphics (COMP 557) Thurs. Feb. 19, 2015 Professor Michael Langer Midterm Exam Fundamentals of Computer Graphics (COMP 557) Thurs. Feb. 19, 2015 Professor Michael Langer The exam consists of 10 questions. There are 2 points per question for a total of 20 points. You

More information

Statistics: Interpreting Data and Making Predictions. Visual Displays of Data 1/31

Statistics: Interpreting Data and Making Predictions. Visual Displays of Data 1/31 Statistics: Interpreting Data and Making Predictions Visual Displays of Data 1/31 Last Time Last time we discussed central tendency; that is, notions of the middle of data. More specifically we discussed

More information

ENCM 501 Winter 2017 Assignment 3 for the Week of January 30

ENCM 501 Winter 2017 Assignment 3 for the Week of January 30 page 1 of 7 ENCM 501 Winter 2017 Assignment 3 for the Week of January 30 Steve Norman Department of Electrical & Computer Engineering University of Calgary January 2017 Assignment instructions and other

More information

CMSC 476/676 Information Retrieval Midterm Exam Spring 2014

CMSC 476/676 Information Retrieval Midterm Exam Spring 2014 CMSC 476/676 Information Retrieval Midterm Exam Spring 2014 Name: You may consult your notes and/or your textbook. This is a 75 minute, in class exam. If there is information missing in any of the question

More information

Variables and Data Representation

Variables and Data Representation You will recall that a computer program is a set of instructions that tell a computer how to transform a given set of input into a specific output. Any program, procedural, event driven or object oriented

More information

CMPT 354 Database Systems I. Spring 2012 Instructor: Hassan Khosravi

CMPT 354 Database Systems I. Spring 2012 Instructor: Hassan Khosravi CMPT 354 Database Systems I Spring 2012 Instructor: Hassan Khosravi Textbook First Course in Database Systems, 3 rd Edition. Jeffry Ullman and Jennifer Widom Other text books Ramakrishnan SILBERSCHATZ

More information

Lecture 27: Learning from relational data

Lecture 27: Learning from relational data Lecture 27: Learning from relational data STATS 202: Data mining and analysis December 2, 2017 1 / 12 Announcements Kaggle deadline is this Thursday (Dec 7) at 4pm. If you haven t already, make a submission

More information

Typing Control. Chapter Conditionals

Typing Control. Chapter Conditionals Chapter 26 Typing Control 26.1 Conditionals Let s expand our language with a conditional construct. We can use if0 like before, but for generality it s going to be more convenient to have a proper conditional

More information

CS101 Introduction to Programming Languages and Compilers

CS101 Introduction to Programming Languages and Compilers CS101 Introduction to Programming Languages and Compilers In this handout we ll examine different types of programming languages and take a brief look at compilers. We ll only hit the major highlights

More information

BTEC Nationals IT - Unit2 FAQs

BTEC Nationals IT - Unit2 FAQs BTEC Nationals IT - Unit2 FAQs Q1 Q2 I need more clarity on what is required in the design task Is it expected that the race officials are entering times as raw times and then the table is set up so it

More information

University of Virginia Department of Computer Science. CS 4501: Information Retrieval Fall 2015

University of Virginia Department of Computer Science. CS 4501: Information Retrieval Fall 2015 University of Virginia Department of Computer Science CS 4501: Information Retrieval Fall 2015 5:00pm-6:15pm, Monday, October 26th Name: ComputingID: This is a closed book and closed notes exam. No electronic

More information

Using Arithmetic of Real Numbers to Explore Limits and Continuity

Using Arithmetic of Real Numbers to Explore Limits and Continuity Using Arithmetic of Real Numbers to Explore Limits and Continuity by Maria Terrell Cornell University Problem Let a =.898989... and b =.000000... (a) Find a + b. (b) Use your ideas about how to add a and

More information

EECS 492 Midterm #1. Example Questions. Note: Not a complete exam!

EECS 492 Midterm #1. Example Questions. Note: Not a complete exam! EECS 492 Midterm #1 Example Questions Note: Not a complete exam! General Instructions This exam is closed book, except that you are allowed to refer to a single sheet of paper. You may use a calculator

More information

DM 505 Database Design and Programming. Spring 2012 Project. Department of Mathematics and Computer Science University of Southern Denmark

DM 505 Database Design and Programming. Spring 2012 Project. Department of Mathematics and Computer Science University of Southern Denmark DM 505 Database Design and Programming Spring 2012 Project Department of Mathematics and Computer Science University of Southern Denmark February 20, 2012 2 Introduction The purpose of this project is

More information

Prepare a stem-and-leaf graph for the following data. In your final display, you should arrange the leaves for each stem in increasing order.

Prepare a stem-and-leaf graph for the following data. In your final display, you should arrange the leaves for each stem in increasing order. Chapter 2 2.1 Descriptive Statistics A stem-and-leaf graph, also called a stemplot, allows for a nice overview of quantitative data without losing information on individual observations. It can be a good

More information

LING 510, Lab 3 September 23, 2013

LING 510, Lab 3 September 23, 2013 LING 510, Lab 3 September 23, 2013 Agenda: Go over Homework 1 Go over JYW, if there are questions Go over function application (what we ended with on Thursday) 1. Frequently missed questions on Homework

More information

THE UNIVERSITY OF BRITISH COLUMBIA CPSC 121: MIDTERM 2 Group March 12, 2014

THE UNIVERSITY OF BRITISH COLUMBIA CPSC 121: MIDTERM 2 Group March 12, 2014 THE UNIVERSITY OF BRITISH COLUMBIA CPSC 121: MIDTERM 2 Group March 12, 2014 Important notes about this examination 1. You have 40 minutes to complete this exam. 2. No electronic aides (e.g., phones or

More information

Exact Inference: Elimination and Sum Product (and hidden Markov models)

Exact Inference: Elimination and Sum Product (and hidden Markov models) Exact Inference: Elimination and Sum Product (and hidden Markov models) David M. Blei Columbia University October 13, 2015 The first sections of these lecture notes follow the ideas in Chapters 3 and 4

More information

CSE 332, Spring 2010, Midterm Examination 30 April 2010

CSE 332, Spring 2010, Midterm Examination 30 April 2010 CSE 332, Spring 2010, Midterm Examination 30 April 2010 Please do not turn the page until the bell rings. Rules: The exam is closed-book, closed-note. You may use a calculator for basic arithmetic only.

More information

ENCM 501 Winter 2015 Assignment 3 for the Week of February 2

ENCM 501 Winter 2015 Assignment 3 for the Week of February 2 page 1 of 6 ENCM 501 Winter 2015 Assignment 3 for the Week of February 2 Steve Norman Department of Electrical & Computer Engineering University of Calgary January 2015 Assignment instructions and other

More information

Lesson 10. Student Outcomes. Lesson Notes

Lesson 10. Student Outcomes. Lesson Notes Student Outcomes Students understand that a function from one set (called the domain) to another set (called the range) assigns each element of the domain to exactly one element of the range and understand

More information

cs173: Programming Languages Midterm Exam

cs173: Programming Languages Midterm Exam cs173: Programming Languages Midterm Exam Fall 2002 Please read this page of instructions before you turn the page! This exam is worth 181 points. We will assign partial credit to partial responses, provided

More information

Shallow Parsing Swapnil Chaudhari 11305R011 Ankur Aher Raj Dabre 11305R001

Shallow Parsing Swapnil Chaudhari 11305R011 Ankur Aher Raj Dabre 11305R001 Shallow Parsing Swapnil Chaudhari 11305R011 Ankur Aher - 113059006 Raj Dabre 11305R001 Purpose of the Seminar To emphasize on the need for Shallow Parsing. To impart basic information about techniques

More information

Probabilistic Graphical Models

Probabilistic Graphical Models 10-708 Probabilistic Graphical Models Homework 4 Due Apr 27, 12:00 noon Submission: Homework is due on the due date at 12:00 noon. Please see course website for policy on late submission. You must submit

More information

CS103 Handout 13 Fall 2012 May 4, 2012 Problem Set 5

CS103 Handout 13 Fall 2012 May 4, 2012 Problem Set 5 CS103 Handout 13 Fall 2012 May 4, 2012 Problem Set 5 This fifth problem set explores the regular languages, their properties, and their limits. This will be your first foray into computability theory,

More information

MIDTERM EXAMINATION Networked Life (NETS 112) November 21, 2013 Prof. Michael Kearns

MIDTERM EXAMINATION Networked Life (NETS 112) November 21, 2013 Prof. Michael Kearns MIDTERM EXAMINATION Networked Life (NETS 112) November 21, 2013 Prof. Michael Kearns This is a closed-book exam. You should have no material on your desk other than the exam itself and a pencil or pen.

More information

Try typing the following in the Python shell and press return after each calculation. Write the answer the program displays next to the sums below.

Try typing the following in the Python shell and press return after each calculation. Write the answer the program displays next to the sums below. Name: Date: Instructions: PYTHON - INTRODUCTORY TASKS Open Idle (the program we will be using to write our Python codes). We can use the following code in Python to work out numeracy calculations. Try

More information

Query Difficulty Prediction for Contextual Image Retrieval

Query Difficulty Prediction for Contextual Image Retrieval Query Difficulty Prediction for Contextual Image Retrieval Xing Xing 1, Yi Zhang 1, and Mei Han 2 1 School of Engineering, UC Santa Cruz, Santa Cruz, CA 95064 2 Google Inc., Mountain View, CA 94043 Abstract.

More information

CS 2506 Computer Organization II Test 1

CS 2506 Computer Organization II Test 1 Instructions: Print your name in the space provided below. This examination is closed book and closed notes, aside from the permitted one-page formula sheet. No calculators or other computing devices may

More information

Lecture 1: Overview

Lecture 1: Overview 15-150 Lecture 1: Overview Lecture by Stefan Muller May 21, 2018 Welcome to 15-150! Today s lecture was an overview that showed the highlights of everything you re learning this semester, which also meant

More information

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu Natural Language Processing CS 6320 Lecture 6 Neural Language Models Instructor: Sanda Harabagiu In this lecture We shall cover: Deep Neural Models for Natural Language Processing Introduce Feed Forward

More information

Range Minimum Queries Part Two

Range Minimum Queries Part Two Range Minimum Queries Part Two Recap from Last Time The RMQ Problem The Range Minimum Query (RMQ) problem is the following: Given a fixed array A and two indices i j, what is the smallest element out of

More information

CIS 121 Data Structures and Algorithms with Java Spring 2018

CIS 121 Data Structures and Algorithms with Java Spring 2018 CIS 121 Data Structures and Algorithms with Java Spring 2018 Homework 6 Compression Due: Monday, March 12, 11:59pm online 2 Required Problems (45 points), Qualitative Questions (10 points), and Style and

More information

Notebook Assignments

Notebook Assignments Notebook Assignments These six assignments are a notebook using techniques from class in the single concrete context of graph theory. This is supplemental to your usual assignments, and is designed for

More information

CSE 131 Introduction to Computer Science Fall Exam I

CSE 131 Introduction to Computer Science Fall Exam I CSE 131 Introduction to Computer Science Fall 2015 Given: 24 September 2015 Exam I Due: End of session This exam is closed-book, closed-notes, no electronic devices allowed. The exception is the sage page

More information

Integrated Math 1 Module 3 Honors Sequences and Series Ready, Set, Go! Homework

Integrated Math 1 Module 3 Honors Sequences and Series Ready, Set, Go! Homework 1 Integrated Math 1 Module 3 Honors Sequences and Series Ready, Set, Go! Homework Adapted from The Mathematics Vision Project: Scott Hendrickson, Joleigh Honey, Barbara Kuehl, Travis Lemon, Janet Sutorius

More information

AP Statistics Summer Math Packet

AP Statistics Summer Math Packet NAME: AP Statistics Summer Math Packet PERIOD: Complete all sections of this packet and bring in with you to turn in on the first day of school. ABOUT THIS SUMMER PACKET: In general, AP Statistics includes

More information

Important Project Dates

Important Project Dates Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.035, Fall 2002 Handout 4 Project Overview Wednesday, September 4 This is an overview of the course project

More information

Math 7 Notes - Unit 4 Pattern & Functions

Math 7 Notes - Unit 4 Pattern & Functions Math 7 Notes - Unit 4 Pattern & Functions Syllabus Objective: (.) The student will create tables, charts, and graphs to etend a pattern in order to describe a linear rule, including integer values. Syllabus

More information

Programming Standards: You must conform to good programming/documentation standards. Some specifics:

Programming Standards: You must conform to good programming/documentation standards. Some specifics: CS3114 (Spring 2011) PROGRAMMING ASSIGNMENT #3 Due Thursday, April 7 @ 11:00 PM for 100 points Early bonus date: Wednesday, April 6 @ 11:00 PM for a 10 point bonus Initial Schedule due Thursday, March

More information

Do not start the test until instructed to do so!

Do not start the test until instructed to do so! CS 3204 Operating Systems Midterm (Abrams) Spring 2004 VIRG INIA POLYTECHNIC INSTITUTE AND STATE U T PRO SI M UNI VERSI TY Instructions: Do not start the test until instructed to do so! Print your name

More information

Industry Competency Examination Proctor Guide

Industry Competency Examination Proctor Guide Industry Competency Examination Proctor Guide TABLE OF CONTENTS INDUSTRY COMPETENCY EXAMINATION...1 INTRODUCTION...1 RECEIPT OF EXAMINATION MATERIALS...1 EXAMINATION SECURITY... 2 EXAMINATION SCHEDULE...

More information

Learning By Sample Series. Learning By Sample - T-SQL : Part 3

Learning By Sample Series. Learning By Sample - T-SQL : Part 3 Learning By Sample Series Learning By Sample - T-SQL : Part 3 w w w. B u k a n S e m b a r a n g. I n f o 2 0 1 0 Foreword Learning By Sample? So, why should I made this tutorial freely accessed by everyone?

More information

Read and fill in this page now. Your instructional login (e.g., cs3-ab): Your lab section days and time: Name of the person sitting to your left:

Read and fill in this page now. Your instructional login (e.g., cs3-ab): Your lab section days and time: Name of the person sitting to your left: CS3 Fall 05 Midterm 1 Read and fill in this page now Your name: Your instructional login (e.g., cs3-ab): Your lab section days and time: Your lab T.A.: Name of the person sitting to your left: Name of

More information

Error Analysis, Statistics and Graphing

Error Analysis, Statistics and Graphing Error Analysis, Statistics and Graphing This semester, most of labs we require us to calculate a numerical answer based on the data we obtain. A hard question to answer in most cases is how good is your

More information

COS 226 Midterm Exam, Spring 2009

COS 226 Midterm Exam, Spring 2009 NAME: login ID: precept: COS 226 Midterm Exam, Spring 2009 This test is 10 questions, weighted as indicated. The exam is closed book, except that you are allowed to use a one page cheatsheet. No calculators

More information

The role of functional decomposition

The role of functional decomposition The role of functional decomposition Jon Bell Doc. ref. SD/TR/FR/10: July 27, 2004 Abstract Hierarchical decomposition of function is already a feature of the language used for interpretation of simulation

More information

Introduction to Sets and Logic (MATH 1190)

Introduction to Sets and Logic (MATH 1190) Introduction to Sets and Logic () Instructor: Email: shenlili@yorku.ca Department of Mathematics and Statistics York University Dec 4, 2014 Outline 1 2 3 4 Definition A relation R from a set A to a set

More information

Software Engineering Prof.N.L.Sarda IIT Bombay. Lecture-11 Data Modelling- ER diagrams, Mapping to relational model (Part -II)

Software Engineering Prof.N.L.Sarda IIT Bombay. Lecture-11 Data Modelling- ER diagrams, Mapping to relational model (Part -II) Software Engineering Prof.N.L.Sarda IIT Bombay Lecture-11 Data Modelling- ER diagrams, Mapping to relational model (Part -II) We will continue our discussion on process modeling. In the previous lecture

More information

1 Counting triangles and cliques

1 Counting triangles and cliques ITCSC-INC Winter School 2015 26 January 2014 notes by Andrej Bogdanov Today we will talk about randomness and some of the surprising roles it plays in the theory of computing and in coding theory. Let

More information

Final Exam. Advanced Methods for Data Analysis (36-402/36-608) Due Thursday May 8, 2014 at 11:59pm

Final Exam. Advanced Methods for Data Analysis (36-402/36-608) Due Thursday May 8, 2014 at 11:59pm Final Exam Advanced Methods for Data Analysis (36-402/36-608) Due Thursday May 8, 2014 at 11:59pm Instructions: you will submit this take-home final exam in three parts. 1. Writeup. This will be a complete

More information

Homework 1 Due Tuesday, January 30, 2018 at 8pm

Homework 1 Due Tuesday, January 30, 2018 at 8pm CSECE 374 A Spring 2018 Homework 1 Due Tuesday, January 30, 2018 at 8pm Starting with this homework, groups of up to three people can submit joint solutions. Each problem should be submitted by exactly

More information

Important Project Dates

Important Project Dates Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.035, Spring 2010 Handout Project Overview Tuesday, Feb 2 This is an overview of the course project and

More information

DOWNLOAD PDF BIG IDEAS MATH VERTICAL SHRINK OF A PARABOLA

DOWNLOAD PDF BIG IDEAS MATH VERTICAL SHRINK OF A PARABOLA Chapter 1 : BioMath: Transformation of Graphs Use the results in part (a) to identify the vertex of the parabola. c. Find a vertical line on your graph paper so that when you fold the paper, the left portion

More information

Fall 2005 Joseph/Tygar/Vazirani/Wagner Final

Fall 2005 Joseph/Tygar/Vazirani/Wagner Final CS 161 Computer Security Fall 2005 Joseph/Tygar/Vazirani/Wagner Final PRINT your name:, (last) SIGN your name: (first) PRINT your Unix account name: PRINT your TA s name: You may consult any books, notes,

More information

Stanford University Computer Science Department Solved CS347 Spring 2001 Mid-term.

Stanford University Computer Science Department Solved CS347 Spring 2001 Mid-term. Stanford University Computer Science Department Solved CS347 Spring 2001 Mid-term. Question 1: (4 points) Shown below is a portion of the positional index in the format term: doc1: position1,position2

More information

Part II Composition of Functions

Part II Composition of Functions Part II Composition of Functions The big idea in this part of the book is deceptively simple. It s that we can take the value returned by one function and use it as an argument to another function. By

More information

Midterm 2 Solutions. CS70 Discrete Mathematics for Computer Science, Fall 2007

Midterm 2 Solutions. CS70 Discrete Mathematics for Computer Science, Fall 2007 CS70 Discrete Mathematics for Computer Science, Fall 007 Midterm Solutions Note: These solutions are not necessarily model answers Rather, they are designed to be tutorial in nature, and sometimes contain

More information

AP Statistics Summer Review Packet

AP Statistics Summer Review Packet 60 NAME: PERIOD: AP Statistics Summer Review Packet Teacher(s): Ryan Oben Teacher(s) Contact Information: Ryan_Oben@mcpsmd.org Course: Purpose of the Summer Assignment: In general, AP Statistics includes

More information

Project 2: Genetic Programming for Symbolic Regression

Project 2: Genetic Programming for Symbolic Regression Project 2: Genetic Programming for Symbolic Regression Assigned: Tuesday, Oct. 2 Multiple Due Dates (all submissions must be submitted via BlackBoard): Undergrads: Generally complete, compilable code:

More information

CS 188: Artificial Intelligence Fall Machine Learning

CS 188: Artificial Intelligence Fall Machine Learning CS 188: Artificial Intelligence Fall 2007 Lecture 23: Naïve Bayes 11/15/2007 Dan Klein UC Berkeley Machine Learning Up till now: how to reason or make decisions using a model Machine learning: how to select

More information

3rd Grade Texas Math Crosswalk Document:

3rd Grade Texas Math Crosswalk Document: New TX Math 3.1(A) apply mathematics to problems arising in everyday life, society, and the workplace; 3.1(B) use a problem-solving model that incorporates analyzing given information, formulating a plan

More information

DEPARTMENT OF ACADEMIC UPGRADING

DEPARTMENT OF ACADEMIC UPGRADING DEPARTMENT OF ACADEMIC UPGRADING COURSE OUTLINE WINTER 2013 INTRODUCTION TO MATH 0081 INSTRUCTOR: Sukhvir Sandhu PHONE: (780) 539-2810 or 2234 OFFICE: Math Lab A210 or C310 E-MAIL: ssandhu@gprc.ab.ca OFFICE

More information

CSE 131 Introduction to Computer Science Fall Final Exam

CSE 131 Introduction to Computer Science Fall Final Exam CSE 131 Introduction to Computer Science Fall 2017 Given: 16 December 2017 Final Exam Due: End of session This exam is closed-book, closed-notes, no electronic devices allowed. The exception is the sage

More information

2. You are required to enter a password of up to 100 characters. The characters must be lower ASCII, printing characters.

2. You are required to enter a password of up to 100 characters. The characters must be lower ASCII, printing characters. BLACK BOX SOFTWARE TESTING SPRING 2005 DOMAIN TESTING LAB PROJECT -- GRADING NOTES For all of the cases below, do the traditional equivalence class and boundary analysis. Draw one table and use a new line

More information

CS 224N: Assignment #1

CS 224N: Assignment #1 Due date: assignment) 1/25 11:59 PM PST (You are allowed to use three (3) late days maximum for this These questions require thought, but do not require long answers. Please be as concise as possible.

More information

Properties and Definitions

Properties and Definitions Section 0.1 Contents: Operations Defined Multiplication as an Abbreviation Visualizing Multiplication Commutative Properties Parentheses Associative Properties Identities Zero Product Answers to Exercises

More information

CS/ECE 374 Fall Homework 1. Due Tuesday, September 6, 2016 at 8pm

CS/ECE 374 Fall Homework 1. Due Tuesday, September 6, 2016 at 8pm CSECE 374 Fall 2016 Homework 1 Due Tuesday, September 6, 2016 at 8pm Starting with this homework, groups of up to three people can submit joint solutions. Each problem should be submitted by exactly one

More information

Finite Element Analysis Prof. Dr. B. N. Rao Department of Civil Engineering Indian Institute of Technology, Madras. Lecture - 24

Finite Element Analysis Prof. Dr. B. N. Rao Department of Civil Engineering Indian Institute of Technology, Madras. Lecture - 24 Finite Element Analysis Prof. Dr. B. N. Rao Department of Civil Engineering Indian Institute of Technology, Madras Lecture - 24 So in today s class, we will look at quadrilateral elements; and we will

More information

ICT WORKSHEET GRADE 9 TERM1 EXAMINATION

ICT WORKSHEET GRADE 9 TERM1 EXAMINATION ICT WORKSHEET GRADE 9 TERM1 EXAMINATION Q1. (A) Give the most suitable device and one advantage and disadvantage of suggested method to (i) Input sound. (ii) Insert a hand-drawn map into a document (iii)

More information

L322 Syntax. Chapter 3: Structural Relations. Linguistics 322 D E F G H. Another representation is in the form of labelled brackets:

L322 Syntax. Chapter 3: Structural Relations. Linguistics 322 D E F G H. Another representation is in the form of labelled brackets: L322 Syntax Chapter 3: Structural Relations Linguistics 322 1 The Parts of a Tree A tree structure is one of an indefinite number of ways to represent a sentence or a part of it. Consider the following

More information

Dependency grammar and dependency parsing

Dependency grammar and dependency parsing Dependency grammar and dependency parsing Syntactic analysis (5LN455) 2014-12-10 Sara Stymne Department of Linguistics and Philology Based on slides from Marco Kuhlmann Mid-course evaluation Mostly positive

More information

Hidden Markov Models. Natural Language Processing: Jordan Boyd-Graber. University of Colorado Boulder LECTURE 20. Adapted from material by Ray Mooney

Hidden Markov Models. Natural Language Processing: Jordan Boyd-Graber. University of Colorado Boulder LECTURE 20. Adapted from material by Ray Mooney Hidden Markov Models Natural Language Processing: Jordan Boyd-Graber University of Colorado Boulder LECTURE 20 Adapted from material by Ray Mooney Natural Language Processing: Jordan Boyd-Graber Boulder

More information

Spring 2007 Midterm Exam

Spring 2007 Midterm Exam 15-381 Spring 2007 Midterm Exam Spring 2007 March 8 Name: Andrew ID: This is an open-book, open-notes examination. You have 80 minutes to complete this examination. Unless explicitly requested, we do not

More information

Lign/CSE 256, Programming Assignment 1: Language Models

Lign/CSE 256, Programming Assignment 1: Language Models Lign/CSE 256, Programming Assignment 1: Language Models 16 January 2008 due 1 Feb 2008 1 Preliminaries First, make sure you can access the course materials. 1 The components are: ˆ code1.zip: the Java

More information

EXCEL BASICS: PROJECTS

EXCEL BASICS: PROJECTS EXCEL BASICS: PROJECTS In this class, you will be practicing with three basic Excel worksheets to learn a variety of foundational skills necessary for more advanced projects. This class covers: Three Project

More information