HW3 Solutions. Answer: Let X be the random variable which denotes the number of packets lost.
|
|
- Cory Walsh
- 5 years ago
- Views:
Transcription
1 HW3 Solutions 1. (20 pts.) Packets Over the Internet n packets are sent over the Internet (n even). Consider the following probability models for the process: (a) Each packet is routed over a different path and is lost independently with probability p. (b) All n packets are routed along the same path and with probability p, one of the links along the path fails and all n packets are lost. Otherwise all packets are received. (c) The n packets are divided into 2 groups of n/2 packets, and each group is routed along a different path and lost with probability p. Losses of different groups are independent events. In each of the three models, compute the distribution of the number of packet losses. For n 6 and p 0.3, plot the distribution in each of the three cases. Does the distribution depend on the probability model? Which of the three routing protocols do you prefer? Answer: Let X be the random variable which denotes the number of packets lost. (a) (6 pts.) Since each of the n packets is independently lost with probability p, we know that X Bin(n, p). Hence, for k {0, 1,..., n}, ( ) n P(X k) p k (1 p) n k k (b) (6 pts.) In this case, either all or none of the packets are lost. So P(X 0) 1 p and P(X n) p. 1
2 (c) (6 pts.) In this case, either no packets, exactly n/2 packets, or all n packets are lost. When n/2 packets are lost, it could be because path 1 failed while path 2 worked, or because path 2 failed while path 1 worked. Let Y i represent whether path i fails (Y i 1 if it fails, Y i 0 if it does not). Then X (n/2)(y 1 + Y 2 ) and Y i Bern(p). Hence, we get P(X 0) (1 p) 2 P(X n/2) 2p(1 p) P(X n) p 2 (2 pts.) As can be seen from the probability distribution tables and plots, the distributions depend on the probability models of packet loss. Which protocol is most preferable is quite dependent on the context, and it s possible to construct situations where each of the protocols performs best: 2
3 If the packets were being transmitted without using an error-correcting code, then all of them are needed at the recipient s end to reconstruct the whole message. In this case, protocol (a) succeeds with probability (1 p) n, protocol (b) with probability 1 p and protocol (c) with probability (1 p) 2. Therefore, protocol (b), which routes all packets on the same path, would work best in this situation. Suppose you used an error-correcting code that can tolerate the loss of up to 35% of packets, and suppose n 100 and p 0.3. Then protocol (a) succeeds with probability 0.884, protocol (b) with probability 0.7, and protocol (c) with probability Therefore, protocol (a) works best in this situation. Suppose you used an error-correcting code that can tolerate the loss of up to 50% of packets, and suppose n 3 and p 0.5. Then protocol (a) and (b) succeed with probability 0.5, while protocol (c) succeeds with probability In this case, protocol (c) works best. Finally, suppose that each packet that makes it to the recipient is worth the same amount; the total value of the communication is proportional to the number of packets received. Then it might not matter which protocol you choose: in the long run, they will deliver the same value, on average. Any of these sorts of answers would be acceptable, as long as you give a coherent justification for which protocol you said is preferable. Comment: There is a particularly important and common scenario where protocol (a) performs best by a wide margin. Suppose we know the packet loss probability p. Then if the number of packets n is large, it turns out that the actual number of packets lost will be pretty close to np (it will almost always be between n(p ɛ) and n(p + ɛ), for some small constant ɛ). Hence, in such a case, it makes sense to use an error-correcting code which is designed to correct up to a p + ɛ of lost packets. By doing so, protocol (a) almost always succeeds in delivering enough packets that the error-correcting code can make up for the ones that were lost. This success probability can be made as close to 1 as desired (by increasing n). On the other hand, protocol (b) will drop all the packets with probability p and protocol (c) with probability p 2, so they are considerably worse in this situation. 2. (20 pts.) Coupon Collection Consider the coupon collection problem with n distinct baseball cards and m cereal boxes bought. Compute the distribution, mean and variance of the number of Babe Ruth cards acquired. (Hint: Think and write in terms of random variables.) Answer: Each box of cereal is analogous to a biased coin flip. We either get Babe Ruth (coin comes up 1 heads) w.p. n or we don t w.p. 1 1 n (coin comes up tails). Since each box of cereal is independent of all others then X Bin(m, 1/n) ( ) ( ) m 1 n 1 m k P(X k) k n k n where X is the random variable denoting the number of cards Babe Ruth cards acquired. 3. (20 pts.) Family Planning Mr and Mrs Brown decide to continue having children until they either have their first girl or until they have five children. Assume that each child is equally likely to be a boy or a girl, independent of all other children, and that there are no multiple births. Let B and G denote the numbers of boys and girls respectively that the Browns have. (a) (10 pts.) Write down the sample space together with the probability of each outcome 3
4 Answer: Ω {g, bg, bbg, bbbg, bbbbg, bbbbb}, where b represents boy and g represents girl in the strings in Ω. For any ω Ω, let l(ω) represent the length of the string (i.e. the number of children born to the Browns). Then, P(ω) ( 1 2 )l(ω), since the gender of all children are independent and each gender is equally likely for every child. (b) (10 pts.) Compute and plot the distributions of the random variables B and G. Answer: P(G 0) P(G 1) The distribution of B is as follows: P(B 0) 1 2 P(B 1) P(B 2) P(B 3) P(B 4) P(B 5)
5 4. (20 pts.) Conditional Independence and Medical Diagnosis There is a disease which affects 2% of the population. A medical test is available which gives a false positive rate of 5% and a mis-detection rate of 3%. (a) (10 pts.) Compute the false discovery rate and the false omission rate of using this test. Answer: Define the random variables H and T as in the lecture notes (H 0 if the patient is healthy, H 1 if affected, T 0 if the test is negative, T 1 if positive). P(H 1) Misdetection rate is P(T 0 H 1) False positive rate is P(T 1 H 0) To compute the false discovery rate, by Bayes rule, P(H 0)P(T 1 H 0) P(H 0 T 1) P(H 0)P(T 1 H 0) + P(H 1)P(T 1 H 1) (1 0.02)0.05 (1 0.02) (1 0.03) To compute the false omission rate, by Bayes rule, P(H 1)P(T 0 H 1) P(H 1 T 0) P(H 1)P(T 0 H 1) + P(H 0)P(T 0 H 0) (1 0.02)(1 0.05) (b) (10 pts.) The doctors complain that the false discovery rate for this test is too high. The medical test company responds with a new test that has a false positive rate of 6% and a mis-detection rate of 4%. Although this new test has worse false positive rate and mis-detection rate compared to the old test, the company claims that when used in conjunction with the old test, will give a lower false discovery 5
6 rate because the results of the two tests are conditionally independent given the disease state of the patient. More specifically, the company recommends the doctors diagnose a patient to have a disease if and only if both the old and the new tests are positive. Do you agree that the company s claim? Justify your answer. Answer: Let the result of the old test be T 1, and that of the new test be T 2. To compute the false discovery rate of the combined test, P(H 0 T 1 1, T 2 1) P(H 0)P(T 1 1, T 2 1 H 0) P(H 0)P(T 1 1, T 2 1 H 0) + P(H 1)P(T 1 1, T 2 1 H 1) P(H 0)P(T 1 1 H 0)P(T 2 1 H 0) P(H 0)P(T 1 1 H 0)P(T 2 1 H 0) + P(H 1)P(T 1 1 H 1)P(T 2 1 H 1) (1 0.02) (1 0.02) (1 0.03)(1 0.04) The false discovery rate is indeed lower. 5. (20 pts.) Typos Assume a 10-page scientific journal paper has 10 typos in it. (a) (10 pts.) If the typos are randomly distributed over the 10 pages, find the probability of having a single typo on each page. Answer: Each typo can be in any of the 10 pages, thus there are 10 different positions for each typo. Since there are 10 pages, the total number of ways we can distribute the typos to the pages is Having a single typo on each of the 10 pages implies sampling the typos 10 times without replacement, since no more than a single typo can be on any page. Thus, there are 10! ways of having a single typo on each page. Hence, the probability of this event is 10!/ (b) (10 pts.) If 8 of the typos are spelling errors and the other two are mathematical errors, find the probability of having 4 spelling errors and a single mathematical error on the first five pages of the paper. Answer: Given a set of 4 spelling errors and 1 mathematical errors, there are 5 5 ways of arranging them over the first five pages, and there are 5 5 ways of arranging the remaining 5 errors over the last 5 pages, leading to a total of 5 10 arrangements. Since there are ( 8 2 4)( 1) different sets of 4 spelling and 1 mathematical errors, the total number of ways we can arrange the errors is ( ( 8 4) 2 1) Hence, the probability of this event is ( ( ( 8 4) 2 1) 5 10 )/
3. Probability 51. probability A numerical value between 0 and 1 assigned to an event to indicate how often the event occurs (in the long run).
3. Probability 51 3 Probability 3.1 Key Definitions and Ideas random process A repeatable process that has multiple unpredictable potential outcomes. Although we sometimes use language that suggests that
More informationDiscrete Mathematics and Probability Theory Fall 2013 Midterm #2
CS 70 Discrete Mathematics and Probability Theory Fall 013 Midterm # 1. (30 points) Short Answer. No justification necessary. Please write only the answer in the space provided after each question. Please
More informationUCSD CSE 21, Spring 2014 [Section B00] Mathematics for Algorithm and System Analysis
UCSD CSE 21, Spring 2014 [Section B00] Mathematics for Algorithm and System Analysis Lecture 11 Class URL: http://vlsicad.ucsd.edu/courses/cse21-s14/ Lecture 11 Notes Goals for this week (Tuesday) Linearity
More informationDiscrete Mathematics Course Review 3
21-228 Discrete Mathematics Course Review 3 This document contains a list of the important definitions and theorems that have been covered thus far in the course. It is not a complete listing of what has
More informationMidterm 2. Friday, August 2, 5:10pm 7:10pm CS 70: Discrete Mathematics and Probability Theory, Summer 2013
Midterm 2 Friday, August 2, 5:10pm 7:10pm CS 70: Discrete Mathematics and Probability Theory, Summer 2013 Your name Your student ID # Your section # This exam has 5 questions and a total of 100 points.
More informationTopic:- DU_J18_MA_STATS_Topic01
DU MA MSc Statistics Topic:- DU_J18_MA_STATS_Topic01 1) In analysis of variance problem involving 3 treatments with 10 observations each, SSE= 399.6. Then the MSE is equal to: [Question ID = 2313] 1. 14.8
More information9.2 Types of Errors in Hypothesis testing
9.2 Types of Errors in Hypothesis testing 1 Mistakes we could make As I mentioned, when we take a sample we won t be 100% sure of something because we do not take a census (we only look at information
More informationLab 3: Building Compound Comparisons
Lab 3: Building Compound Comparisons In this lab you will build a series of Compound Comparisons. Each Compound Comparison will relate to a group of related Identifiers. And each will hold an ordered list
More information15-780: Problem Set #3
15-780: Problem Set #3 March 31, 2014 1. Partial-Order Planning [15 pts] Consider a version of the milk//banana//drill shopping problem used in the second planning lecture (slides 17 26). (a) Modify the
More informationChapter 3. Set Theory. 3.1 What is a Set?
Chapter 3 Set Theory 3.1 What is a Set? A set is a well-defined collection of objects called elements or members of the set. Here, well-defined means accurately and unambiguously stated or described. Any
More information1 Probability Review. CS 124 Section #8 Hashing, Skip Lists 3/20/17. Expectation (weighted average): the expectation of a random quantity X is:
CS 124 Section #8 Hashing, Skip Lists 3/20/17 1 Probability Review Expectation (weighted average): the expectation of a random quantity X is: x= x P (X = x) For each value x that X can take on, we look
More informationWe will show that the height of a RB tree on n vertices is approximately 2*log n. In class I presented a simple structural proof of this claim:
We have seen that the insert operation on a RB takes an amount of time proportional to the number of the levels of the tree (since the additional operations required to do any rebalancing require constant
More informationGraph Contraction. Graph Contraction CSE341T/CSE549T 10/20/2014. Lecture 14
CSE341T/CSE549T 10/20/2014 Lecture 14 Graph Contraction Graph Contraction So far we have mostly talking about standard techniques for solving problems on graphs that were developed in the context of sequential
More informationBayes Theorem. Theorem: Let {A i } n i=1
Bayes Theorem Theorem: Let {A i } n i=1 be a set of n events, each with positive probability that partition S in such a way that n i=1 A i = S, and A i A j = φ, i j. For any event B with P(B) > 0, for
More informationLecture 3: Conditional Independence - Undirected
CS598: Graphical Models, Fall 2016 Lecture 3: Conditional Independence - Undirected Lecturer: Sanmi Koyejo Scribe: Nate Bowman and Erin Carrier, Aug. 30, 2016 1 Review for the Bayes-Ball Algorithm Recall
More information1 Leaffix Scan, Rootfix Scan, Tree Size, and Depth
Lecture 17 Graph Contraction I: Tree Contraction Parallel and Sequential Data Structures and Algorithms, 15-210 (Spring 2012) Lectured by Kanat Tangwongsan March 20, 2012 In this lecture, we will explore
More information1 (15 points) LexicoSort
CS161 Homework 2 Due: 22 April 2016, 12 noon Submit on Gradescope Handed out: 15 April 2016 Instructions: Please answer the following questions to the best of your ability. If you are asked to show your
More informationNaïve Bayes Classification. Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others
Naïve Bayes Classification Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others Things We d Like to Do Spam Classification Given an email, predict
More informationNaïve Bayes Classification. Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others
Naïve Bayes Classification Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others Things We d Like to Do Spam Classification Given an email, predict
More informationMATHEMATICS 191, FALL 2004 MATHEMATICAL PROBABILITY Outline #1 (Countability and Uncountability)
MATHEMATICS 191, FALL 2004 MATHEMATICAL PROBABILITY Outline #1 (Countability and Uncountability) Last modified: September 16, 2004 Reference: Apostol, Calculus, Vol. 2, section 13.19 (attached). The aim
More informationWeek - 01 Lecture - 03 Euclid's Algorithm for gcd. Let us continue with our running example of gcd to explore more issues involved with program.
Programming, Data Structures and Algorithms in Python Prof. Madhavan Mukund Department of Computer Science and Engineering Indian Institute of Technology, Madras Week - 01 Lecture - 03 Euclid's Algorithm
More informationWeek 2: Frequency distributions
Types of data Health Sciences M.Sc. Programme Applied Biostatistics Week 2: distributions Data can be summarised to help to reveal information they contain. We do this by calculating numbers from the data
More informationPROVIDER WEBSITE SITE ADMINISTRATOR GUIDE » PATIENT INQUIRY» CLAIM CENTER» FIND A DOCTOR» CLAIMS EDITING SYSTEM (CES)
PROVIDER WEBSITE SITE ADMINISTRATOR GUIDE» PATIENT INQUIRY» CLAIM CENTER» FIND A DOCTOR» CLAIMS EDITING SYSTEM (CES) 2018 WPS Health Plan, Inc. 1 All rights reserved. JO7048 28898-085-1801 ADMINISTRATIVE
More informationCMPSCI611: Approximating SET-COVER Lecture 21
CMPSCI611: Approximating SET-COVER Lecture 21 Today we look at two more examples of approximation algorithms for NP-hard optimization problems. The first, for the SET-COVER problem, has an approximation
More information23.2 Normal Distributions
1_ Locker LESSON 23.2 Normal Distributions Common Core Math Standards The student is expected to: S-ID.4 Use the mean and standard deviation of a data set to fit it to a normal distribution and to estimate
More informationBinary Decision Diagrams
Logic and roof Hilary 2016 James Worrell Binary Decision Diagrams A propositional formula is determined up to logical equivalence by its truth table. If the formula has n variables then its truth table
More informationCSE 417 Dynamic Programming (pt 6) Parsing Algorithms
CSE 417 Dynamic Programming (pt 6) Parsing Algorithms Reminders > HW9 due on Friday start early program will be slow, so debugging will be slow... should run in 2-4 minutes > Please fill out course evaluations
More informationMidterm 2 Solutions. CS70 Discrete Mathematics for Computer Science, Fall 2007
CS70 Discrete Mathematics for Computer Science, Fall 007 Midterm Solutions Note: These solutions are not necessarily model answers Rather, they are designed to be tutorial in nature, and sometimes contain
More informationUNIVERSITY OF EDINBURGH COLLEGE OF SCIENCE AND ENGINEERING SCHOOL OF INFORMATICS INFR08008 INFORMATICS 2A: PROCESSING FORMAL AND NATURAL LANGUAGES
UNIVERSITY OF EDINBURGH COLLEGE OF SCIENCE AND ENGINEERING SCHOOL OF INFORMATICS INFR08008 INFORMATICS 2A: PROCESSING FORMAL AND NATURAL LANGUAGES Saturday 10 th December 2016 09:30 to 11:30 INSTRUCTIONS
More informationCPSC 536N: Randomized Algorithms Term 2. Lecture 5
CPSC 536N: Randomized Algorithms 2011-12 Term 2 Prof. Nick Harvey Lecture 5 University of British Columbia In this lecture we continue to discuss applications of randomized algorithms in computer networking.
More informationMA 1128: Lecture 02 1/22/2018
MA 1128: Lecture 02 1/22/2018 Exponents Scientific Notation 1 Exponents Exponents are used to indicate how many copies of a number are to be multiplied together. For example, I like to deal with the signs
More informationStatistics 13, Lab 1. Getting Started. The Mac. Launching RStudio and loading data
Statistics 13, Lab 1 Getting Started This first lab session is nothing more than an introduction: We will help you navigate the Statistics Department s (all Mac) computing facility and we will get you
More informationCSCE 478/878 Lecture 6: Bayesian Learning and Graphical Models. Stephen Scott. Introduction. Outline. Bayes Theorem. Formulas
ian ian ian Might have reasons (domain information) to favor some hypotheses/predictions over others a priori ian methods work with probabilities, and have two main roles: Optimal Naïve Nets (Adapted from
More informationDiscrete Math: Selected Homework Problems
Discrete Math: Selected Homework Problems 2006 2.1 Prove: if d is a common divisor of a and b and d is also a linear combination of a and b then d is a greatest common divisor of a and b. (5 3.1 Prove:
More informationMore About WHILE Loops
More About WHILE Loops http://people.sc.fsu.edu/ jburkardt/isc/week04 lecture 07.pdf... ISC3313: Introduction to Scientific Computing with C++ Summer Semester 2011... John Burkardt Department of Scientific
More informationMassachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Recitation-6: Hardness of Inference Contents 1 NP-Hardness Part-II
More informationMidterm 2 Solutions. CS70 Discrete Mathematics and Probability Theory, Spring 2009
CS70 Discrete Mathematics and Probability Theory, Spring 2009 Midterm 2 Solutions Note: These solutions are not necessarily model answers. Rather, they are designed to be tutorial in nature, and sometimes
More informationPrivacy in Statistical Databases
Privacy in Statistical Databases CSE 598D/STAT 598B Fall 2007 Lecture 2, 9/13/2007 Aleksandra Slavkovic Office hours: MW 3:30-4:30 Office: Thomas 412 Phone: x3-4918 Adam Smith Office hours: Mondays 3-5pm
More information/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18
601.433/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18 22.1 Introduction We spent the last two lectures proving that for certain problems, we can
More informationEarthquake data in geonet.org.nz
Earthquake data in geonet.org.nz There is are large gaps in the 2012 and 2013 data, so let s not use it. Instead we ll use a previous year. Go to http://http://quakesearch.geonet.org.nz/ At the screen,
More informationExercises with switch
These are all short programs. None of these programs should exceed 30 lines. A line means any expression ending with a ;, not including header files. Some of these are as short as 10 lines. Some problems
More informationComputational Optimization ISE 407. Lecture 16. Dr. Ted Ralphs
Computational Optimization ISE 407 Lecture 16 Dr. Ted Ralphs ISE 407 Lecture 16 1 References for Today s Lecture Required reading Sections 6.5-6.7 References CLRS Chapter 22 R. Sedgewick, Algorithms in
More information6c Lecture 3 & 4: April 8 & 10, 2014
6c Lecture 3 & 4: April 8 & 10, 2014 3.1 Graphs and trees We begin by recalling some basic definitions from graph theory. Definition 3.1. A (undirected, simple) graph consists of a set of vertices V and
More informationIn our first lecture on sets and set theory, we introduced a bunch of new symbols and terminology.
Guide to and Hi everybody! In our first lecture on sets and set theory, we introduced a bunch of new symbols and terminology. This guide focuses on two of those symbols: and. These symbols represent concepts
More informationBig-O-ology 1 CIS 675: Algorithms January 14, 2019
Big-O-ology 1 CIS 675: Algorithms January 14, 2019 1. The problem Consider a carpenter who is building you an addition to your house. You would not think much of this carpenter if she or he couldn t produce
More informationLecture 3: Linear Classification
Lecture 3: Linear Classification Roger Grosse 1 Introduction Last week, we saw an example of a learning task called regression. There, the goal was to predict a scalar-valued target from a set of features.
More informationAssignment 4 CSE 517: Natural Language Processing
Assignment 4 CSE 517: Natural Language Processing University of Washington Winter 2016 Due: March 2, 2016, 1:30 pm 1 HMMs and PCFGs Here s the definition of a PCFG given in class on 2/17: A finite set
More informationPRODUCT DOCUMENTATION. Association Discovery
PRODUCT DOCUMENTATION Association Discovery BigML Documentation: Association Discovery!2 BigML Inc. 2851 NW 9th, Corvallis, OR 97330, U.S. https://bigml.com info@bigml.com BigML Documentation: Association
More informationList of Exercises: Data Mining 1 December 12th, 2015
List of Exercises: Data Mining 1 December 12th, 2015 1. We trained a model on a two-class balanced dataset using five-fold cross validation. One person calculated the performance of the classifier by measuring
More informationGLOSSARY OF TERMS. Commutative property. Numbers can be added or multiplied in either order. For example, = ; 3 x 8 = 8 x 3.
GLOSSARY OF TERMS Algorithm. An established step-by-step procedure used 1 to achieve a desired result. For example, the 55 addition algorithm for the sum of two two-digit + 27 numbers where carrying is
More informationWarm-up for Integrated Geometry and Algebra I
Summer Assignment Warm-up for Integrated Geometry and Algebra I Who should complete this packet? Students who will be taking Integrated Geometry and Algebra I in the fall of 018. Due Date: The first day
More informationThe Basics of Graphical Models
The Basics of Graphical Models David M. Blei Columbia University September 30, 2016 1 Introduction (These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan.
More informationBinary Search Trees, etc.
Chapter 12 Binary Search Trees, etc. Binary Search trees are data structures that support a variety of dynamic set operations, e.g., Search, Minimum, Maximum, Predecessors, Successors, Insert, and Delete.
More informationDiscrete Mathematics and Probability Theory Summer 2016 Dinh, Psomas, and Ye HW 2
CS 70 Discrete Mathematics and Probability Theory Summer 2016 Dinh, Psomas, and Ye HW 2 Due Tuesday July 5 at 1:59PM 1. (8 points: 3/5) Hit or miss For each of the claims and proofs below, state whether
More informationx y
10. LECTURE 10 Objectives I understand the difficulty in finding an appropriate function for a data set in general. In some cases, I can define a function type that may fit a data set well. Last time,
More informationUniversität Freiburg Lehrstuhl für Maschinelles Lernen und natürlichsprachliche Systeme. Machine Learning (SS2012)
Universität Freiburg Lehrstuhl für Maschinelles Lernen und natürlichsprachliche Systeme Machine Learning (SS2012) Prof. Dr. M. Riedmiller, Manuel Blum Exercise Sheet 5 Exercise 5.1: Spam Detection Suppose
More informationCOMP3121/3821/9101/ s1 Assignment 1
Sample solutions to assignment 1 1. (a) Describe an O(n log n) algorithm (in the sense of the worst case performance) that, given an array S of n integers and another integer x, determines whether or not
More informationGraphical Models. David M. Blei Columbia University. September 17, 2014
Graphical Models David M. Blei Columbia University September 17, 2014 These lecture notes follow the ideas in Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. In addition,
More informationIn this packet, you will find the forms necessary for your child to apply to the Registry. To apply, please complete the following steps:
Dear Registry Applicant, Thank you for your interest in the National Registry of Myotonic Dystrophy and Facioscapulohumeral Muscular Dystrophy Patients and Family Members! The Registry was established
More informationCSE 21 Mathematics for Algorithm and System Analysis
CSE 21 Mathematics for Algorithm and System Analysis Unit 4: Basic Concepts in Graph Theory Section 3: Trees 1 Review : Decision Tree (DT-Section 1) Root of the decision tree on the left: 1 Leaves of the
More informationParameter Estimation. Learning From Data: MLE. Parameter Estimation. Likelihood. Maximum Likelihood Parameter Estimation. Likelihood Function 12/1/16
Learning From Data: MLE Maximum Estimators Common approach in statistics: use a parametric model of data: Assume data set: Bin(n, p), Poisson( ), N(µ, exp( ) Uniform(a, b) 2 ) But parameters are unknown!!!
More informationWELCOME! Lecture 3 Thommy Perlinger
Quantitative Methods II WELCOME! Lecture 3 Thommy Perlinger Program Lecture 3 Cleaning and transforming data Graphical examination of the data Missing Values Graphical examination of the data It is important
More informationRunning Minitab for the first time on your PC
Running Minitab for the first time on your PC Screen Appearance When you select the MINITAB option from the MINITAB 14 program group, or click on MINITAB 14 under RAS you will see the following screen.
More informationAI Programming CS S-15 Probability Theory
AI Programming CS662-2013S-15 Probability Theory David Galles Department of Computer Science University of San Francisco 15-0: Uncertainty In many interesting agent environments, uncertainty plays a central
More informationMath 120 Introduction to Statistics Mr. Toner s Lecture Notes 3.1 Measures of Central Tendency
Math 1 Introduction to Statistics Mr. Toner s Lecture Notes 3.1 Measures of Central Tendency lowest value + highest value midrange The word average: is very ambiguous and can actually refer to the mean,
More informationSubmit: Your group source code to mooshak
Tutorial 2 (Optional) Genetic Algorithms This is an optional tutorial. Should you decide to answer it please Submit: Your group source code to mooshak http://mooshak.deei.fct.ualg.pt/ up to May 28, 2018.
More informationRecall from last time. Lecture 4: Wrap-up of Bayes net representation. Markov networks. Markov blanket. Isolating a node
Recall from last time Lecture 4: Wrap-up of Bayes net representation. Markov networks Markov blanket, moral graph Independence maps and perfect maps Undirected graphical models (Markov networks) A Bayes
More information6.867 Machine Learning
6.867 Machine Learning Problem set - solutions Thursday, October What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove. Do not
More informationMath 101 Final Exam Study Notes:
Math 101 Final Exam Study Notes: *Please remember there is a large set of final exam review problems in Doc Sharing (under Course Tools in MLP). Highlighted are what might be considered formulas* I. Graph
More informationEvaluating Machine-Learning Methods. Goals for the lecture
Evaluating Machine-Learning Methods Mark Craven and David Page Computer Sciences 760 Spring 2018 www.biostat.wisc.edu/~craven/cs760/ Some of the slides in these lectures have been adapted/borrowed from
More informationCondence Intervals about a Single Parameter:
Chapter 9 Condence Intervals about a Single Parameter: 9.1 About a Population Mean, known Denition 9.1.1 A point estimate of a parameter is the value of a statistic that estimates the value of the parameter.
More information1KOd17RMoURxjn2 CSE 20 DISCRETE MATH Fall
CSE 20 https://goo.gl/forms/1o 1KOd17RMoURxjn2 DISCRETE MATH Fall 2017 http://cseweb.ucsd.edu/classes/fa17/cse20-ab/ Today's learning goals Explain the steps in a proof by mathematical and/or structural
More informationThings you ll know (or know better to watch out for!) when you leave in December: 1. What you can and cannot infer from graphs.
1 2 Things you ll know (or know better to watch out for!) when you leave in December: 1. What you can and cannot infer from graphs. 2. How to construct (in your head!) and interpret confidence intervals.
More information1 More configuration model
1 More configuration model In the last lecture, we explored the definition of the configuration model, a simple method for drawing networks from the ensemble, and derived some of its mathematical properties.
More informationPASS Sample Size Software. Randomization Lists
Chapter 880 Introduction This module is used to create a randomization list for assigning subjects to one of up to eight groups or treatments. Six randomization algorithms are available. Four of the algorithms
More informationAckworth Howard Church of England (VC) Junior and Infant School. Child-friendly GDPR privacy notice
Child-friendly GDPR privacy notice Child-friendly GDPR privacy notice What s this about? A new law has been made that keeps your information safe things like your address, date of birth and phone number.
More informationCS369G: Algorithmic Techniques for Big Data Spring
CS369G: Algorithmic Techniques for Big Data Spring 2015-2016 Lecture 11: l 0 -Sampling and Introduction to Graph Streaming Prof. Moses Charikar Scribe: Austin Benson 1 Overview We present and analyze the
More informationCPSC 121: Models of Computation Assignment #4, due Thursday, March 16 th,2017at16:00
CPSC 121: Models of Computation Assignment #4, due Thursday, March 16 th,2017at16:00 [18] 1. Consider the following predicate logic statement: 9x 2 A, 8y 2 B,9z 2 C, P(x, y, z)! Q(x, y, z), where A, B,
More informationCS264: Homework #4. Due by midnight on Wednesday, October 22, 2014
CS264: Homework #4 Due by midnight on Wednesday, October 22, 2014 Instructions: (1) Form a group of 1-3 students. You should turn in only one write-up for your entire group. (2) Turn in your solutions
More information2. A Bernoulli distribution has the following likelihood function for a data set D: N 1 N 1 + N 0
Machine Learning Fall 2015 Homework 1 Homework must be submitted electronically following the instructions on the course homepage. Make sure to explain you reasoning or show your derivations. Except for
More informationCOT 3100 Spring 2010 Midterm 2
COT 3100 Spring 2010 Midterm 2 For the first two questions #1 and #2, do ONLY ONE of them. If you do both, only question #1 will be graded. 1. (20 pts) Give iterative and recursive algorithms for finding
More information/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Priority Queues / Heaps Date: 9/27/17
01.433/33 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Priority Queues / Heaps Date: 9/2/1.1 Introduction In this lecture we ll talk about a useful abstraction, priority queues, which are
More informationCS6375: Machine Learning Gautam Kunapuli. Mid-Term Review
Gautam Kunapuli Machine Learning Data is identically and independently distributed Goal is to learn a function that maps to Data is generated using an unknown function Learn a hypothesis that minimizes
More informationPart II. Graph Theory. Year
Part II Year 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2017 53 Paper 3, Section II 15H Define the Ramsey numbers R(s, t) for integers s, t 2. Show that R(s, t) exists for all s,
More informationCIS192 Python Programming
CIS192 Python Programming Probability, Simulations & Course Wrap-Up Raymond Yin University of Pennsylvania December 8, 2016 Raymond Yin (University of Pennsylvania) CIS 192 December 8, 2016 1 / 21 Outline
More informationMarch 13, Bellwork. Due Today!
Bellwork Due Today! Bellwork March 13, 2014 Bellwork Kayla has taken three tests worth 100 points each. Her scores are 85, 93, and 88. She has one test left to take. What score must she get on her last
More informationCS 308 Data Structures Spring Dr. George Bebis Programming Assignment 3
CS 308 Data Structures Spring 2003 - Dr. George Bebis Programming Assignment 3 Due date:04/15/03 In this assignment, you will use linked-lists to store appropriate information (i.e., centroid, size, pixels
More informationThe Probabilistic Method
The Probabilistic Method Po-Shen Loh June 2010 1 Warm-up 1. (Russia 1996/4 In the Duma there are 1600 delegates, who have formed 16000 committees of 80 persons each. Prove that one can find two committees
More informationData Mining Classification: Bayesian Decision Theory
Data Mining Classification: Bayesian Decision Theory Lecture Notes for Chapter 2 R. O. Duda, P. E. Hart, and D. G. Stork, Pattern classification, 2nd ed. New York: Wiley, 2001. Lecture Notes for Chapter
More informationComputational Complexity and Implications for Security DRAFT Notes on Infeasible Computation for MA/CS 109 Leo Reyzin with the help of Nick Benes
Computational Complexity and Implications for Security DRAFT Notes on Infeasible Computation for MA/CS 109 Leo Reyzin with the help of Nick Benes The Study of Computational Complexity Let s summarize what
More informationTracking Database. COL live (COL only)
II. Tracking Database The Tracking Database is used to keep track of information about each participant enrolled in the COL/AS+ Project. Native Boys and Girls Club staff who are participating in the project
More informationRadix Searching. The insert procedure for digital search trees also derives directly from the corresponding procedure for binary search trees:
Radix Searching The most simple radix search method is digital tree searching - the binary search tree with the branch in the tree according to the bits of keys: at the first level the leading bit is used,
More informationMODIFIED VERSION OF: An introduction to Matlab for dynamic modeling ***PART 3 ***
MODIFIED VERSION OF: An introduction to Matlab for dynamic modeling ***PART 3 *** Stephen P. Ellner 1 and John Guckenheimer 2 1 Department of Ecology and Evolutionary Biology, and 2 Department of Mathematics
More informationDO NOT REPRODUCE. CS61B, Fall 2008 Test #3 (revised) P. N. Hilfinger
CS6B, Fall 2008 Test #3 (revised) P. N. Hilfinger. [7 points] Please give short answers to the following, giving reasons where called for. Unless a question says otherwise, time estimates refer to asymptotic
More informationEvaluating Machine Learning Methods: Part 1
Evaluating Machine Learning Methods: Part 1 CS 760@UW-Madison Goals for the lecture you should understand the following concepts bias of an estimator learning curves stratified sampling cross validation
More informationPractice Exam III - Answers
Department of Mathematics University of Notre Dame Math 10120 Finite Math Spring 2015 Name: Instructor: Migliore Practice Exam III - Answers April 16, 2015 This exam is in two parts on 13 pages and contains
More informationSIDM3: Combining and restructuring datasets; creating summary data across repeated measures or across groups
SIDM3: Combining and restructuring datasets; creating summary data across repeated measures or across groups You might find that your data is in a very different structure to that needed for analysis.
More informationlecture notes September 2, How to sort?
.30 lecture notes September 2, 203 How to sort? Lecturer: Michel Goemans The task of sorting. Setup Suppose we have n objects that we need to sort according to some ordering. These could be integers or
More informationCS 2316 Exam 4 Fall 2011
CS 2316 Exam 4 Fall 2011 Name : Grading TA: Integrity: By taking this exam, you pledge that this is your work and you have neither given nor received inappropriate help during the taking of this exam in
More informationHomework 1. Due Date: Wednesday 11/26/07 - at the beginning of the lecture
Homework 1 Due Date: Wednesday 11/26/07 - at the beginning of the lecture Problems marked with a [*] are a littlebit harder and count as extra credit. Note 1. For any of the given problems make sure that
More information