HW3 Solutions. Answer: Let X be the random variable which denotes the number of packets lost.

Size: px
Start display at page:

Download "HW3 Solutions. Answer: Let X be the random variable which denotes the number of packets lost."

Transcription

1 HW3 Solutions 1. (20 pts.) Packets Over the Internet n packets are sent over the Internet (n even). Consider the following probability models for the process: (a) Each packet is routed over a different path and is lost independently with probability p. (b) All n packets are routed along the same path and with probability p, one of the links along the path fails and all n packets are lost. Otherwise all packets are received. (c) The n packets are divided into 2 groups of n/2 packets, and each group is routed along a different path and lost with probability p. Losses of different groups are independent events. In each of the three models, compute the distribution of the number of packet losses. For n 6 and p 0.3, plot the distribution in each of the three cases. Does the distribution depend on the probability model? Which of the three routing protocols do you prefer? Answer: Let X be the random variable which denotes the number of packets lost. (a) (6 pts.) Since each of the n packets is independently lost with probability p, we know that X Bin(n, p). Hence, for k {0, 1,..., n}, ( ) n P(X k) p k (1 p) n k k (b) (6 pts.) In this case, either all or none of the packets are lost. So P(X 0) 1 p and P(X n) p. 1

2 (c) (6 pts.) In this case, either no packets, exactly n/2 packets, or all n packets are lost. When n/2 packets are lost, it could be because path 1 failed while path 2 worked, or because path 2 failed while path 1 worked. Let Y i represent whether path i fails (Y i 1 if it fails, Y i 0 if it does not). Then X (n/2)(y 1 + Y 2 ) and Y i Bern(p). Hence, we get P(X 0) (1 p) 2 P(X n/2) 2p(1 p) P(X n) p 2 (2 pts.) As can be seen from the probability distribution tables and plots, the distributions depend on the probability models of packet loss. Which protocol is most preferable is quite dependent on the context, and it s possible to construct situations where each of the protocols performs best: 2

3 If the packets were being transmitted without using an error-correcting code, then all of them are needed at the recipient s end to reconstruct the whole message. In this case, protocol (a) succeeds with probability (1 p) n, protocol (b) with probability 1 p and protocol (c) with probability (1 p) 2. Therefore, protocol (b), which routes all packets on the same path, would work best in this situation. Suppose you used an error-correcting code that can tolerate the loss of up to 35% of packets, and suppose n 100 and p 0.3. Then protocol (a) succeeds with probability 0.884, protocol (b) with probability 0.7, and protocol (c) with probability Therefore, protocol (a) works best in this situation. Suppose you used an error-correcting code that can tolerate the loss of up to 50% of packets, and suppose n 3 and p 0.5. Then protocol (a) and (b) succeed with probability 0.5, while protocol (c) succeeds with probability In this case, protocol (c) works best. Finally, suppose that each packet that makes it to the recipient is worth the same amount; the total value of the communication is proportional to the number of packets received. Then it might not matter which protocol you choose: in the long run, they will deliver the same value, on average. Any of these sorts of answers would be acceptable, as long as you give a coherent justification for which protocol you said is preferable. Comment: There is a particularly important and common scenario where protocol (a) performs best by a wide margin. Suppose we know the packet loss probability p. Then if the number of packets n is large, it turns out that the actual number of packets lost will be pretty close to np (it will almost always be between n(p ɛ) and n(p + ɛ), for some small constant ɛ). Hence, in such a case, it makes sense to use an error-correcting code which is designed to correct up to a p + ɛ of lost packets. By doing so, protocol (a) almost always succeeds in delivering enough packets that the error-correcting code can make up for the ones that were lost. This success probability can be made as close to 1 as desired (by increasing n). On the other hand, protocol (b) will drop all the packets with probability p and protocol (c) with probability p 2, so they are considerably worse in this situation. 2. (20 pts.) Coupon Collection Consider the coupon collection problem with n distinct baseball cards and m cereal boxes bought. Compute the distribution, mean and variance of the number of Babe Ruth cards acquired. (Hint: Think and write in terms of random variables.) Answer: Each box of cereal is analogous to a biased coin flip. We either get Babe Ruth (coin comes up 1 heads) w.p. n or we don t w.p. 1 1 n (coin comes up tails). Since each box of cereal is independent of all others then X Bin(m, 1/n) ( ) ( ) m 1 n 1 m k P(X k) k n k n where X is the random variable denoting the number of cards Babe Ruth cards acquired. 3. (20 pts.) Family Planning Mr and Mrs Brown decide to continue having children until they either have their first girl or until they have five children. Assume that each child is equally likely to be a boy or a girl, independent of all other children, and that there are no multiple births. Let B and G denote the numbers of boys and girls respectively that the Browns have. (a) (10 pts.) Write down the sample space together with the probability of each outcome 3

4 Answer: Ω {g, bg, bbg, bbbg, bbbbg, bbbbb}, where b represents boy and g represents girl in the strings in Ω. For any ω Ω, let l(ω) represent the length of the string (i.e. the number of children born to the Browns). Then, P(ω) ( 1 2 )l(ω), since the gender of all children are independent and each gender is equally likely for every child. (b) (10 pts.) Compute and plot the distributions of the random variables B and G. Answer: P(G 0) P(G 1) The distribution of B is as follows: P(B 0) 1 2 P(B 1) P(B 2) P(B 3) P(B 4) P(B 5)

5 4. (20 pts.) Conditional Independence and Medical Diagnosis There is a disease which affects 2% of the population. A medical test is available which gives a false positive rate of 5% and a mis-detection rate of 3%. (a) (10 pts.) Compute the false discovery rate and the false omission rate of using this test. Answer: Define the random variables H and T as in the lecture notes (H 0 if the patient is healthy, H 1 if affected, T 0 if the test is negative, T 1 if positive). P(H 1) Misdetection rate is P(T 0 H 1) False positive rate is P(T 1 H 0) To compute the false discovery rate, by Bayes rule, P(H 0)P(T 1 H 0) P(H 0 T 1) P(H 0)P(T 1 H 0) + P(H 1)P(T 1 H 1) (1 0.02)0.05 (1 0.02) (1 0.03) To compute the false omission rate, by Bayes rule, P(H 1)P(T 0 H 1) P(H 1 T 0) P(H 1)P(T 0 H 1) + P(H 0)P(T 0 H 0) (1 0.02)(1 0.05) (b) (10 pts.) The doctors complain that the false discovery rate for this test is too high. The medical test company responds with a new test that has a false positive rate of 6% and a mis-detection rate of 4%. Although this new test has worse false positive rate and mis-detection rate compared to the old test, the company claims that when used in conjunction with the old test, will give a lower false discovery 5

6 rate because the results of the two tests are conditionally independent given the disease state of the patient. More specifically, the company recommends the doctors diagnose a patient to have a disease if and only if both the old and the new tests are positive. Do you agree that the company s claim? Justify your answer. Answer: Let the result of the old test be T 1, and that of the new test be T 2. To compute the false discovery rate of the combined test, P(H 0 T 1 1, T 2 1) P(H 0)P(T 1 1, T 2 1 H 0) P(H 0)P(T 1 1, T 2 1 H 0) + P(H 1)P(T 1 1, T 2 1 H 1) P(H 0)P(T 1 1 H 0)P(T 2 1 H 0) P(H 0)P(T 1 1 H 0)P(T 2 1 H 0) + P(H 1)P(T 1 1 H 1)P(T 2 1 H 1) (1 0.02) (1 0.02) (1 0.03)(1 0.04) The false discovery rate is indeed lower. 5. (20 pts.) Typos Assume a 10-page scientific journal paper has 10 typos in it. (a) (10 pts.) If the typos are randomly distributed over the 10 pages, find the probability of having a single typo on each page. Answer: Each typo can be in any of the 10 pages, thus there are 10 different positions for each typo. Since there are 10 pages, the total number of ways we can distribute the typos to the pages is Having a single typo on each of the 10 pages implies sampling the typos 10 times without replacement, since no more than a single typo can be on any page. Thus, there are 10! ways of having a single typo on each page. Hence, the probability of this event is 10!/ (b) (10 pts.) If 8 of the typos are spelling errors and the other two are mathematical errors, find the probability of having 4 spelling errors and a single mathematical error on the first five pages of the paper. Answer: Given a set of 4 spelling errors and 1 mathematical errors, there are 5 5 ways of arranging them over the first five pages, and there are 5 5 ways of arranging the remaining 5 errors over the last 5 pages, leading to a total of 5 10 arrangements. Since there are ( 8 2 4)( 1) different sets of 4 spelling and 1 mathematical errors, the total number of ways we can arrange the errors is ( ( 8 4) 2 1) Hence, the probability of this event is ( ( ( 8 4) 2 1) 5 10 )/

3. Probability 51. probability A numerical value between 0 and 1 assigned to an event to indicate how often the event occurs (in the long run).

3. Probability 51. probability A numerical value between 0 and 1 assigned to an event to indicate how often the event occurs (in the long run). 3. Probability 51 3 Probability 3.1 Key Definitions and Ideas random process A repeatable process that has multiple unpredictable potential outcomes. Although we sometimes use language that suggests that

More information

Discrete Mathematics and Probability Theory Fall 2013 Midterm #2

Discrete Mathematics and Probability Theory Fall 2013 Midterm #2 CS 70 Discrete Mathematics and Probability Theory Fall 013 Midterm # 1. (30 points) Short Answer. No justification necessary. Please write only the answer in the space provided after each question. Please

More information

UCSD CSE 21, Spring 2014 [Section B00] Mathematics for Algorithm and System Analysis

UCSD CSE 21, Spring 2014 [Section B00] Mathematics for Algorithm and System Analysis UCSD CSE 21, Spring 2014 [Section B00] Mathematics for Algorithm and System Analysis Lecture 11 Class URL: http://vlsicad.ucsd.edu/courses/cse21-s14/ Lecture 11 Notes Goals for this week (Tuesday) Linearity

More information

Discrete Mathematics Course Review 3

Discrete Mathematics Course Review 3 21-228 Discrete Mathematics Course Review 3 This document contains a list of the important definitions and theorems that have been covered thus far in the course. It is not a complete listing of what has

More information

Midterm 2. Friday, August 2, 5:10pm 7:10pm CS 70: Discrete Mathematics and Probability Theory, Summer 2013

Midterm 2. Friday, August 2, 5:10pm 7:10pm CS 70: Discrete Mathematics and Probability Theory, Summer 2013 Midterm 2 Friday, August 2, 5:10pm 7:10pm CS 70: Discrete Mathematics and Probability Theory, Summer 2013 Your name Your student ID # Your section # This exam has 5 questions and a total of 100 points.

More information

Topic:- DU_J18_MA_STATS_Topic01

Topic:- DU_J18_MA_STATS_Topic01 DU MA MSc Statistics Topic:- DU_J18_MA_STATS_Topic01 1) In analysis of variance problem involving 3 treatments with 10 observations each, SSE= 399.6. Then the MSE is equal to: [Question ID = 2313] 1. 14.8

More information

9.2 Types of Errors in Hypothesis testing

9.2 Types of Errors in Hypothesis testing 9.2 Types of Errors in Hypothesis testing 1 Mistakes we could make As I mentioned, when we take a sample we won t be 100% sure of something because we do not take a census (we only look at information

More information

Lab 3: Building Compound Comparisons

Lab 3: Building Compound Comparisons Lab 3: Building Compound Comparisons In this lab you will build a series of Compound Comparisons. Each Compound Comparison will relate to a group of related Identifiers. And each will hold an ordered list

More information

15-780: Problem Set #3

15-780: Problem Set #3 15-780: Problem Set #3 March 31, 2014 1. Partial-Order Planning [15 pts] Consider a version of the milk//banana//drill shopping problem used in the second planning lecture (slides 17 26). (a) Modify the

More information

Chapter 3. Set Theory. 3.1 What is a Set?

Chapter 3. Set Theory. 3.1 What is a Set? Chapter 3 Set Theory 3.1 What is a Set? A set is a well-defined collection of objects called elements or members of the set. Here, well-defined means accurately and unambiguously stated or described. Any

More information

1 Probability Review. CS 124 Section #8 Hashing, Skip Lists 3/20/17. Expectation (weighted average): the expectation of a random quantity X is:

1 Probability Review. CS 124 Section #8 Hashing, Skip Lists 3/20/17. Expectation (weighted average): the expectation of a random quantity X is: CS 124 Section #8 Hashing, Skip Lists 3/20/17 1 Probability Review Expectation (weighted average): the expectation of a random quantity X is: x= x P (X = x) For each value x that X can take on, we look

More information

We will show that the height of a RB tree on n vertices is approximately 2*log n. In class I presented a simple structural proof of this claim:

We will show that the height of a RB tree on n vertices is approximately 2*log n. In class I presented a simple structural proof of this claim: We have seen that the insert operation on a RB takes an amount of time proportional to the number of the levels of the tree (since the additional operations required to do any rebalancing require constant

More information

Graph Contraction. Graph Contraction CSE341T/CSE549T 10/20/2014. Lecture 14

Graph Contraction. Graph Contraction CSE341T/CSE549T 10/20/2014. Lecture 14 CSE341T/CSE549T 10/20/2014 Lecture 14 Graph Contraction Graph Contraction So far we have mostly talking about standard techniques for solving problems on graphs that were developed in the context of sequential

More information

Bayes Theorem. Theorem: Let {A i } n i=1

Bayes Theorem. Theorem: Let {A i } n i=1 Bayes Theorem Theorem: Let {A i } n i=1 be a set of n events, each with positive probability that partition S in such a way that n i=1 A i = S, and A i A j = φ, i j. For any event B with P(B) > 0, for

More information

Lecture 3: Conditional Independence - Undirected

Lecture 3: Conditional Independence - Undirected CS598: Graphical Models, Fall 2016 Lecture 3: Conditional Independence - Undirected Lecturer: Sanmi Koyejo Scribe: Nate Bowman and Erin Carrier, Aug. 30, 2016 1 Review for the Bayes-Ball Algorithm Recall

More information

1 Leaffix Scan, Rootfix Scan, Tree Size, and Depth

1 Leaffix Scan, Rootfix Scan, Tree Size, and Depth Lecture 17 Graph Contraction I: Tree Contraction Parallel and Sequential Data Structures and Algorithms, 15-210 (Spring 2012) Lectured by Kanat Tangwongsan March 20, 2012 In this lecture, we will explore

More information

1 (15 points) LexicoSort

1 (15 points) LexicoSort CS161 Homework 2 Due: 22 April 2016, 12 noon Submit on Gradescope Handed out: 15 April 2016 Instructions: Please answer the following questions to the best of your ability. If you are asked to show your

More information

Naïve Bayes Classification. Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others

Naïve Bayes Classification. Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others Naïve Bayes Classification Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others Things We d Like to Do Spam Classification Given an email, predict

More information

Naïve Bayes Classification. Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others

Naïve Bayes Classification. Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others Naïve Bayes Classification Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others Things We d Like to Do Spam Classification Given an email, predict

More information

MATHEMATICS 191, FALL 2004 MATHEMATICAL PROBABILITY Outline #1 (Countability and Uncountability)

MATHEMATICS 191, FALL 2004 MATHEMATICAL PROBABILITY Outline #1 (Countability and Uncountability) MATHEMATICS 191, FALL 2004 MATHEMATICAL PROBABILITY Outline #1 (Countability and Uncountability) Last modified: September 16, 2004 Reference: Apostol, Calculus, Vol. 2, section 13.19 (attached). The aim

More information

Week - 01 Lecture - 03 Euclid's Algorithm for gcd. Let us continue with our running example of gcd to explore more issues involved with program.

Week - 01 Lecture - 03 Euclid's Algorithm for gcd. Let us continue with our running example of gcd to explore more issues involved with program. Programming, Data Structures and Algorithms in Python Prof. Madhavan Mukund Department of Computer Science and Engineering Indian Institute of Technology, Madras Week - 01 Lecture - 03 Euclid's Algorithm

More information

Week 2: Frequency distributions

Week 2: Frequency distributions Types of data Health Sciences M.Sc. Programme Applied Biostatistics Week 2: distributions Data can be summarised to help to reveal information they contain. We do this by calculating numbers from the data

More information

PROVIDER WEBSITE SITE ADMINISTRATOR GUIDE » PATIENT INQUIRY» CLAIM CENTER» FIND A DOCTOR» CLAIMS EDITING SYSTEM (CES)

PROVIDER WEBSITE SITE ADMINISTRATOR GUIDE » PATIENT INQUIRY» CLAIM CENTER» FIND A DOCTOR» CLAIMS EDITING SYSTEM (CES) PROVIDER WEBSITE SITE ADMINISTRATOR GUIDE» PATIENT INQUIRY» CLAIM CENTER» FIND A DOCTOR» CLAIMS EDITING SYSTEM (CES) 2018 WPS Health Plan, Inc. 1 All rights reserved. JO7048 28898-085-1801 ADMINISTRATIVE

More information

CMPSCI611: Approximating SET-COVER Lecture 21

CMPSCI611: Approximating SET-COVER Lecture 21 CMPSCI611: Approximating SET-COVER Lecture 21 Today we look at two more examples of approximation algorithms for NP-hard optimization problems. The first, for the SET-COVER problem, has an approximation

More information

23.2 Normal Distributions

23.2 Normal Distributions 1_ Locker LESSON 23.2 Normal Distributions Common Core Math Standards The student is expected to: S-ID.4 Use the mean and standard deviation of a data set to fit it to a normal distribution and to estimate

More information

Binary Decision Diagrams

Binary Decision Diagrams Logic and roof Hilary 2016 James Worrell Binary Decision Diagrams A propositional formula is determined up to logical equivalence by its truth table. If the formula has n variables then its truth table

More information

CSE 417 Dynamic Programming (pt 6) Parsing Algorithms

CSE 417 Dynamic Programming (pt 6) Parsing Algorithms CSE 417 Dynamic Programming (pt 6) Parsing Algorithms Reminders > HW9 due on Friday start early program will be slow, so debugging will be slow... should run in 2-4 minutes > Please fill out course evaluations

More information

Midterm 2 Solutions. CS70 Discrete Mathematics for Computer Science, Fall 2007

Midterm 2 Solutions. CS70 Discrete Mathematics for Computer Science, Fall 2007 CS70 Discrete Mathematics for Computer Science, Fall 007 Midterm Solutions Note: These solutions are not necessarily model answers Rather, they are designed to be tutorial in nature, and sometimes contain

More information

UNIVERSITY OF EDINBURGH COLLEGE OF SCIENCE AND ENGINEERING SCHOOL OF INFORMATICS INFR08008 INFORMATICS 2A: PROCESSING FORMAL AND NATURAL LANGUAGES

UNIVERSITY OF EDINBURGH COLLEGE OF SCIENCE AND ENGINEERING SCHOOL OF INFORMATICS INFR08008 INFORMATICS 2A: PROCESSING FORMAL AND NATURAL LANGUAGES UNIVERSITY OF EDINBURGH COLLEGE OF SCIENCE AND ENGINEERING SCHOOL OF INFORMATICS INFR08008 INFORMATICS 2A: PROCESSING FORMAL AND NATURAL LANGUAGES Saturday 10 th December 2016 09:30 to 11:30 INSTRUCTIONS

More information

CPSC 536N: Randomized Algorithms Term 2. Lecture 5

CPSC 536N: Randomized Algorithms Term 2. Lecture 5 CPSC 536N: Randomized Algorithms 2011-12 Term 2 Prof. Nick Harvey Lecture 5 University of British Columbia In this lecture we continue to discuss applications of randomized algorithms in computer networking.

More information

MA 1128: Lecture 02 1/22/2018

MA 1128: Lecture 02 1/22/2018 MA 1128: Lecture 02 1/22/2018 Exponents Scientific Notation 1 Exponents Exponents are used to indicate how many copies of a number are to be multiplied together. For example, I like to deal with the signs

More information

Statistics 13, Lab 1. Getting Started. The Mac. Launching RStudio and loading data

Statistics 13, Lab 1. Getting Started. The Mac. Launching RStudio and loading data Statistics 13, Lab 1 Getting Started This first lab session is nothing more than an introduction: We will help you navigate the Statistics Department s (all Mac) computing facility and we will get you

More information

CSCE 478/878 Lecture 6: Bayesian Learning and Graphical Models. Stephen Scott. Introduction. Outline. Bayes Theorem. Formulas

CSCE 478/878 Lecture 6: Bayesian Learning and Graphical Models. Stephen Scott. Introduction. Outline. Bayes Theorem. Formulas ian ian ian Might have reasons (domain information) to favor some hypotheses/predictions over others a priori ian methods work with probabilities, and have two main roles: Optimal Naïve Nets (Adapted from

More information

Discrete Math: Selected Homework Problems

Discrete Math: Selected Homework Problems Discrete Math: Selected Homework Problems 2006 2.1 Prove: if d is a common divisor of a and b and d is also a linear combination of a and b then d is a greatest common divisor of a and b. (5 3.1 Prove:

More information

More About WHILE Loops

More About WHILE Loops More About WHILE Loops http://people.sc.fsu.edu/ jburkardt/isc/week04 lecture 07.pdf... ISC3313: Introduction to Scientific Computing with C++ Summer Semester 2011... John Burkardt Department of Scientific

More information

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Recitation-6: Hardness of Inference Contents 1 NP-Hardness Part-II

More information

Midterm 2 Solutions. CS70 Discrete Mathematics and Probability Theory, Spring 2009

Midterm 2 Solutions. CS70 Discrete Mathematics and Probability Theory, Spring 2009 CS70 Discrete Mathematics and Probability Theory, Spring 2009 Midterm 2 Solutions Note: These solutions are not necessarily model answers. Rather, they are designed to be tutorial in nature, and sometimes

More information

Privacy in Statistical Databases

Privacy in Statistical Databases Privacy in Statistical Databases CSE 598D/STAT 598B Fall 2007 Lecture 2, 9/13/2007 Aleksandra Slavkovic Office hours: MW 3:30-4:30 Office: Thomas 412 Phone: x3-4918 Adam Smith Office hours: Mondays 3-5pm

More information

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18 601.433/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18 22.1 Introduction We spent the last two lectures proving that for certain problems, we can

More information

Earthquake data in geonet.org.nz

Earthquake data in geonet.org.nz Earthquake data in geonet.org.nz There is are large gaps in the 2012 and 2013 data, so let s not use it. Instead we ll use a previous year. Go to http://http://quakesearch.geonet.org.nz/ At the screen,

More information

Exercises with switch

Exercises with switch These are all short programs. None of these programs should exceed 30 lines. A line means any expression ending with a ;, not including header files. Some of these are as short as 10 lines. Some problems

More information

Computational Optimization ISE 407. Lecture 16. Dr. Ted Ralphs

Computational Optimization ISE 407. Lecture 16. Dr. Ted Ralphs Computational Optimization ISE 407 Lecture 16 Dr. Ted Ralphs ISE 407 Lecture 16 1 References for Today s Lecture Required reading Sections 6.5-6.7 References CLRS Chapter 22 R. Sedgewick, Algorithms in

More information

6c Lecture 3 & 4: April 8 & 10, 2014

6c Lecture 3 & 4: April 8 & 10, 2014 6c Lecture 3 & 4: April 8 & 10, 2014 3.1 Graphs and trees We begin by recalling some basic definitions from graph theory. Definition 3.1. A (undirected, simple) graph consists of a set of vertices V and

More information

In our first lecture on sets and set theory, we introduced a bunch of new symbols and terminology.

In our first lecture on sets and set theory, we introduced a bunch of new symbols and terminology. Guide to and Hi everybody! In our first lecture on sets and set theory, we introduced a bunch of new symbols and terminology. This guide focuses on two of those symbols: and. These symbols represent concepts

More information

Big-O-ology 1 CIS 675: Algorithms January 14, 2019

Big-O-ology 1 CIS 675: Algorithms January 14, 2019 Big-O-ology 1 CIS 675: Algorithms January 14, 2019 1. The problem Consider a carpenter who is building you an addition to your house. You would not think much of this carpenter if she or he couldn t produce

More information

Lecture 3: Linear Classification

Lecture 3: Linear Classification Lecture 3: Linear Classification Roger Grosse 1 Introduction Last week, we saw an example of a learning task called regression. There, the goal was to predict a scalar-valued target from a set of features.

More information

Assignment 4 CSE 517: Natural Language Processing

Assignment 4 CSE 517: Natural Language Processing Assignment 4 CSE 517: Natural Language Processing University of Washington Winter 2016 Due: March 2, 2016, 1:30 pm 1 HMMs and PCFGs Here s the definition of a PCFG given in class on 2/17: A finite set

More information

PRODUCT DOCUMENTATION. Association Discovery

PRODUCT DOCUMENTATION. Association Discovery PRODUCT DOCUMENTATION Association Discovery BigML Documentation: Association Discovery!2 BigML Inc. 2851 NW 9th, Corvallis, OR 97330, U.S. https://bigml.com info@bigml.com BigML Documentation: Association

More information

List of Exercises: Data Mining 1 December 12th, 2015

List of Exercises: Data Mining 1 December 12th, 2015 List of Exercises: Data Mining 1 December 12th, 2015 1. We trained a model on a two-class balanced dataset using five-fold cross validation. One person calculated the performance of the classifier by measuring

More information

GLOSSARY OF TERMS. Commutative property. Numbers can be added or multiplied in either order. For example, = ; 3 x 8 = 8 x 3.

GLOSSARY OF TERMS. Commutative property. Numbers can be added or multiplied in either order. For example, = ; 3 x 8 = 8 x 3. GLOSSARY OF TERMS Algorithm. An established step-by-step procedure used 1 to achieve a desired result. For example, the 55 addition algorithm for the sum of two two-digit + 27 numbers where carrying is

More information

Warm-up for Integrated Geometry and Algebra I

Warm-up for Integrated Geometry and Algebra I Summer Assignment Warm-up for Integrated Geometry and Algebra I Who should complete this packet? Students who will be taking Integrated Geometry and Algebra I in the fall of 018. Due Date: The first day

More information

The Basics of Graphical Models

The Basics of Graphical Models The Basics of Graphical Models David M. Blei Columbia University September 30, 2016 1 Introduction (These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan.

More information

Binary Search Trees, etc.

Binary Search Trees, etc. Chapter 12 Binary Search Trees, etc. Binary Search trees are data structures that support a variety of dynamic set operations, e.g., Search, Minimum, Maximum, Predecessors, Successors, Insert, and Delete.

More information

Discrete Mathematics and Probability Theory Summer 2016 Dinh, Psomas, and Ye HW 2

Discrete Mathematics and Probability Theory Summer 2016 Dinh, Psomas, and Ye HW 2 CS 70 Discrete Mathematics and Probability Theory Summer 2016 Dinh, Psomas, and Ye HW 2 Due Tuesday July 5 at 1:59PM 1. (8 points: 3/5) Hit or miss For each of the claims and proofs below, state whether

More information

x y

x y 10. LECTURE 10 Objectives I understand the difficulty in finding an appropriate function for a data set in general. In some cases, I can define a function type that may fit a data set well. Last time,

More information

Universität Freiburg Lehrstuhl für Maschinelles Lernen und natürlichsprachliche Systeme. Machine Learning (SS2012)

Universität Freiburg Lehrstuhl für Maschinelles Lernen und natürlichsprachliche Systeme. Machine Learning (SS2012) Universität Freiburg Lehrstuhl für Maschinelles Lernen und natürlichsprachliche Systeme Machine Learning (SS2012) Prof. Dr. M. Riedmiller, Manuel Blum Exercise Sheet 5 Exercise 5.1: Spam Detection Suppose

More information

COMP3121/3821/9101/ s1 Assignment 1

COMP3121/3821/9101/ s1 Assignment 1 Sample solutions to assignment 1 1. (a) Describe an O(n log n) algorithm (in the sense of the worst case performance) that, given an array S of n integers and another integer x, determines whether or not

More information

Graphical Models. David M. Blei Columbia University. September 17, 2014

Graphical Models. David M. Blei Columbia University. September 17, 2014 Graphical Models David M. Blei Columbia University September 17, 2014 These lecture notes follow the ideas in Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. In addition,

More information

In this packet, you will find the forms necessary for your child to apply to the Registry. To apply, please complete the following steps:

In this packet, you will find the forms necessary for your child to apply to the Registry. To apply, please complete the following steps: Dear Registry Applicant, Thank you for your interest in the National Registry of Myotonic Dystrophy and Facioscapulohumeral Muscular Dystrophy Patients and Family Members! The Registry was established

More information

CSE 21 Mathematics for Algorithm and System Analysis

CSE 21 Mathematics for Algorithm and System Analysis CSE 21 Mathematics for Algorithm and System Analysis Unit 4: Basic Concepts in Graph Theory Section 3: Trees 1 Review : Decision Tree (DT-Section 1) Root of the decision tree on the left: 1 Leaves of the

More information

Parameter Estimation. Learning From Data: MLE. Parameter Estimation. Likelihood. Maximum Likelihood Parameter Estimation. Likelihood Function 12/1/16

Parameter Estimation. Learning From Data: MLE. Parameter Estimation. Likelihood. Maximum Likelihood Parameter Estimation. Likelihood Function 12/1/16 Learning From Data: MLE Maximum Estimators Common approach in statistics: use a parametric model of data: Assume data set: Bin(n, p), Poisson( ), N(µ, exp( ) Uniform(a, b) 2 ) But parameters are unknown!!!

More information

WELCOME! Lecture 3 Thommy Perlinger

WELCOME! Lecture 3 Thommy Perlinger Quantitative Methods II WELCOME! Lecture 3 Thommy Perlinger Program Lecture 3 Cleaning and transforming data Graphical examination of the data Missing Values Graphical examination of the data It is important

More information

Running Minitab for the first time on your PC

Running Minitab for the first time on your PC Running Minitab for the first time on your PC Screen Appearance When you select the MINITAB option from the MINITAB 14 program group, or click on MINITAB 14 under RAS you will see the following screen.

More information

AI Programming CS S-15 Probability Theory

AI Programming CS S-15 Probability Theory AI Programming CS662-2013S-15 Probability Theory David Galles Department of Computer Science University of San Francisco 15-0: Uncertainty In many interesting agent environments, uncertainty plays a central

More information

Math 120 Introduction to Statistics Mr. Toner s Lecture Notes 3.1 Measures of Central Tendency

Math 120 Introduction to Statistics Mr. Toner s Lecture Notes 3.1 Measures of Central Tendency Math 1 Introduction to Statistics Mr. Toner s Lecture Notes 3.1 Measures of Central Tendency lowest value + highest value midrange The word average: is very ambiguous and can actually refer to the mean,

More information

Submit: Your group source code to mooshak

Submit: Your group source code to mooshak Tutorial 2 (Optional) Genetic Algorithms This is an optional tutorial. Should you decide to answer it please Submit: Your group source code to mooshak http://mooshak.deei.fct.ualg.pt/ up to May 28, 2018.

More information

Recall from last time. Lecture 4: Wrap-up of Bayes net representation. Markov networks. Markov blanket. Isolating a node

Recall from last time. Lecture 4: Wrap-up of Bayes net representation. Markov networks. Markov blanket. Isolating a node Recall from last time Lecture 4: Wrap-up of Bayes net representation. Markov networks Markov blanket, moral graph Independence maps and perfect maps Undirected graphical models (Markov networks) A Bayes

More information

6.867 Machine Learning

6.867 Machine Learning 6.867 Machine Learning Problem set - solutions Thursday, October What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove. Do not

More information

Math 101 Final Exam Study Notes:

Math 101 Final Exam Study Notes: Math 101 Final Exam Study Notes: *Please remember there is a large set of final exam review problems in Doc Sharing (under Course Tools in MLP). Highlighted are what might be considered formulas* I. Graph

More information

Evaluating Machine-Learning Methods. Goals for the lecture

Evaluating Machine-Learning Methods. Goals for the lecture Evaluating Machine-Learning Methods Mark Craven and David Page Computer Sciences 760 Spring 2018 www.biostat.wisc.edu/~craven/cs760/ Some of the slides in these lectures have been adapted/borrowed from

More information

Condence Intervals about a Single Parameter:

Condence Intervals about a Single Parameter: Chapter 9 Condence Intervals about a Single Parameter: 9.1 About a Population Mean, known Denition 9.1.1 A point estimate of a parameter is the value of a statistic that estimates the value of the parameter.

More information

1KOd17RMoURxjn2 CSE 20 DISCRETE MATH Fall

1KOd17RMoURxjn2 CSE 20 DISCRETE MATH Fall CSE 20 https://goo.gl/forms/1o 1KOd17RMoURxjn2 DISCRETE MATH Fall 2017 http://cseweb.ucsd.edu/classes/fa17/cse20-ab/ Today's learning goals Explain the steps in a proof by mathematical and/or structural

More information

Things you ll know (or know better to watch out for!) when you leave in December: 1. What you can and cannot infer from graphs.

Things you ll know (or know better to watch out for!) when you leave in December: 1. What you can and cannot infer from graphs. 1 2 Things you ll know (or know better to watch out for!) when you leave in December: 1. What you can and cannot infer from graphs. 2. How to construct (in your head!) and interpret confidence intervals.

More information

1 More configuration model

1 More configuration model 1 More configuration model In the last lecture, we explored the definition of the configuration model, a simple method for drawing networks from the ensemble, and derived some of its mathematical properties.

More information

PASS Sample Size Software. Randomization Lists

PASS Sample Size Software. Randomization Lists Chapter 880 Introduction This module is used to create a randomization list for assigning subjects to one of up to eight groups or treatments. Six randomization algorithms are available. Four of the algorithms

More information

Ackworth Howard Church of England (VC) Junior and Infant School. Child-friendly GDPR privacy notice

Ackworth Howard Church of England (VC) Junior and Infant School. Child-friendly GDPR privacy notice Child-friendly GDPR privacy notice Child-friendly GDPR privacy notice What s this about? A new law has been made that keeps your information safe things like your address, date of birth and phone number.

More information

CS369G: Algorithmic Techniques for Big Data Spring

CS369G: Algorithmic Techniques for Big Data Spring CS369G: Algorithmic Techniques for Big Data Spring 2015-2016 Lecture 11: l 0 -Sampling and Introduction to Graph Streaming Prof. Moses Charikar Scribe: Austin Benson 1 Overview We present and analyze the

More information

CPSC 121: Models of Computation Assignment #4, due Thursday, March 16 th,2017at16:00

CPSC 121: Models of Computation Assignment #4, due Thursday, March 16 th,2017at16:00 CPSC 121: Models of Computation Assignment #4, due Thursday, March 16 th,2017at16:00 [18] 1. Consider the following predicate logic statement: 9x 2 A, 8y 2 B,9z 2 C, P(x, y, z)! Q(x, y, z), where A, B,

More information

CS264: Homework #4. Due by midnight on Wednesday, October 22, 2014

CS264: Homework #4. Due by midnight on Wednesday, October 22, 2014 CS264: Homework #4 Due by midnight on Wednesday, October 22, 2014 Instructions: (1) Form a group of 1-3 students. You should turn in only one write-up for your entire group. (2) Turn in your solutions

More information

2. A Bernoulli distribution has the following likelihood function for a data set D: N 1 N 1 + N 0

2. A Bernoulli distribution has the following likelihood function for a data set D: N 1 N 1 + N 0 Machine Learning Fall 2015 Homework 1 Homework must be submitted electronically following the instructions on the course homepage. Make sure to explain you reasoning or show your derivations. Except for

More information

COT 3100 Spring 2010 Midterm 2

COT 3100 Spring 2010 Midterm 2 COT 3100 Spring 2010 Midterm 2 For the first two questions #1 and #2, do ONLY ONE of them. If you do both, only question #1 will be graded. 1. (20 pts) Give iterative and recursive algorithms for finding

More information

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Priority Queues / Heaps Date: 9/27/17

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Priority Queues / Heaps Date: 9/27/17 01.433/33 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Priority Queues / Heaps Date: 9/2/1.1 Introduction In this lecture we ll talk about a useful abstraction, priority queues, which are

More information

CS6375: Machine Learning Gautam Kunapuli. Mid-Term Review

CS6375: Machine Learning Gautam Kunapuli. Mid-Term Review Gautam Kunapuli Machine Learning Data is identically and independently distributed Goal is to learn a function that maps to Data is generated using an unknown function Learn a hypothesis that minimizes

More information

Part II. Graph Theory. Year

Part II. Graph Theory. Year Part II Year 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2017 53 Paper 3, Section II 15H Define the Ramsey numbers R(s, t) for integers s, t 2. Show that R(s, t) exists for all s,

More information

CIS192 Python Programming

CIS192 Python Programming CIS192 Python Programming Probability, Simulations & Course Wrap-Up Raymond Yin University of Pennsylvania December 8, 2016 Raymond Yin (University of Pennsylvania) CIS 192 December 8, 2016 1 / 21 Outline

More information

March 13, Bellwork. Due Today!

March 13, Bellwork. Due Today! Bellwork Due Today! Bellwork March 13, 2014 Bellwork Kayla has taken three tests worth 100 points each. Her scores are 85, 93, and 88. She has one test left to take. What score must she get on her last

More information

CS 308 Data Structures Spring Dr. George Bebis Programming Assignment 3

CS 308 Data Structures Spring Dr. George Bebis Programming Assignment 3 CS 308 Data Structures Spring 2003 - Dr. George Bebis Programming Assignment 3 Due date:04/15/03 In this assignment, you will use linked-lists to store appropriate information (i.e., centroid, size, pixels

More information

The Probabilistic Method

The Probabilistic Method The Probabilistic Method Po-Shen Loh June 2010 1 Warm-up 1. (Russia 1996/4 In the Duma there are 1600 delegates, who have formed 16000 committees of 80 persons each. Prove that one can find two committees

More information

Data Mining Classification: Bayesian Decision Theory

Data Mining Classification: Bayesian Decision Theory Data Mining Classification: Bayesian Decision Theory Lecture Notes for Chapter 2 R. O. Duda, P. E. Hart, and D. G. Stork, Pattern classification, 2nd ed. New York: Wiley, 2001. Lecture Notes for Chapter

More information

Computational Complexity and Implications for Security DRAFT Notes on Infeasible Computation for MA/CS 109 Leo Reyzin with the help of Nick Benes

Computational Complexity and Implications for Security DRAFT Notes on Infeasible Computation for MA/CS 109 Leo Reyzin with the help of Nick Benes Computational Complexity and Implications for Security DRAFT Notes on Infeasible Computation for MA/CS 109 Leo Reyzin with the help of Nick Benes The Study of Computational Complexity Let s summarize what

More information

Tracking Database. COL live (COL only)

Tracking Database. COL live (COL only) II. Tracking Database The Tracking Database is used to keep track of information about each participant enrolled in the COL/AS+ Project. Native Boys and Girls Club staff who are participating in the project

More information

Radix Searching. The insert procedure for digital search trees also derives directly from the corresponding procedure for binary search trees:

Radix Searching. The insert procedure for digital search trees also derives directly from the corresponding procedure for binary search trees: Radix Searching The most simple radix search method is digital tree searching - the binary search tree with the branch in the tree according to the bits of keys: at the first level the leading bit is used,

More information

MODIFIED VERSION OF: An introduction to Matlab for dynamic modeling ***PART 3 ***

MODIFIED VERSION OF: An introduction to Matlab for dynamic modeling ***PART 3 *** MODIFIED VERSION OF: An introduction to Matlab for dynamic modeling ***PART 3 *** Stephen P. Ellner 1 and John Guckenheimer 2 1 Department of Ecology and Evolutionary Biology, and 2 Department of Mathematics

More information

DO NOT REPRODUCE. CS61B, Fall 2008 Test #3 (revised) P. N. Hilfinger

DO NOT REPRODUCE. CS61B, Fall 2008 Test #3 (revised) P. N. Hilfinger CS6B, Fall 2008 Test #3 (revised) P. N. Hilfinger. [7 points] Please give short answers to the following, giving reasons where called for. Unless a question says otherwise, time estimates refer to asymptotic

More information

Evaluating Machine Learning Methods: Part 1

Evaluating Machine Learning Methods: Part 1 Evaluating Machine Learning Methods: Part 1 CS 760@UW-Madison Goals for the lecture you should understand the following concepts bias of an estimator learning curves stratified sampling cross validation

More information

Practice Exam III - Answers

Practice Exam III - Answers Department of Mathematics University of Notre Dame Math 10120 Finite Math Spring 2015 Name: Instructor: Migliore Practice Exam III - Answers April 16, 2015 This exam is in two parts on 13 pages and contains

More information

SIDM3: Combining and restructuring datasets; creating summary data across repeated measures or across groups

SIDM3: Combining and restructuring datasets; creating summary data across repeated measures or across groups SIDM3: Combining and restructuring datasets; creating summary data across repeated measures or across groups You might find that your data is in a very different structure to that needed for analysis.

More information

lecture notes September 2, How to sort?

lecture notes September 2, How to sort? .30 lecture notes September 2, 203 How to sort? Lecturer: Michel Goemans The task of sorting. Setup Suppose we have n objects that we need to sort according to some ordering. These could be integers or

More information

CS 2316 Exam 4 Fall 2011

CS 2316 Exam 4 Fall 2011 CS 2316 Exam 4 Fall 2011 Name : Grading TA: Integrity: By taking this exam, you pledge that this is your work and you have neither given nor received inappropriate help during the taking of this exam in

More information

Homework 1. Due Date: Wednesday 11/26/07 - at the beginning of the lecture

Homework 1. Due Date: Wednesday 11/26/07 - at the beginning of the lecture Homework 1 Due Date: Wednesday 11/26/07 - at the beginning of the lecture Problems marked with a [*] are a littlebit harder and count as extra credit. Note 1. For any of the given problems make sure that

More information