A Quantitative Measure of Relevance Based on Kelly Gambling Theory

Size: px
Start display at page:

Download "A Quantitative Measure of Relevance Based on Kelly Gambling Theory"

Transcription

1 A Quantitative Measure of Relevance Based on Kelly Gambling Theory Mathias Winther Madsen Institute for Logic, Language, and Computation University of Amsterdam

2 PLAN Why How Examples

3 Why

4 Why

5 How

6 Why not use Shannon information H(X) == E log Pr(X == x) Claude Shannon (96 200)

7 Why not use Shannon information Information Content === Prior Uncertainty (cf. Klir 2008; Shannon 948) Posterior Uncertainty

8 Why not use Shannon information Pr(X == ) == 0.5 Pr(X == 2) == 0.9 What is the value of X Pr(X == 3) == 0.23 Pr(X == 4) == 0.2 H(X) == E log Pr(X == x) Pr(X == 5) == 0.22 == 2.3

9 Why not use Shannon information 0 Pr(X == ) == 0.5 Pr(X == 2) == 0.9 Is X == 2 0 Is X == 3 0 Pr(X == 3) == 0.23 Is X in {4,5} Is X == 5 Expected number of questions: == 2.34 Pr(X == 4) == Pr(X == 5) == 0.22

10 What color are my socks H(p) == p log p == 6.53 bits of entropy.

11 How

12 Why not use value-of-information $! Value-ofInformation! === $$ $ $$ Posterior Expectation Prior Expectation

13 Why not use value-of-information Rules: Your capital can be distributed freely Bets on the actual outcome are returned twofold Bets on all other outcomes are lost

14 Why not use value-of-information Expected payoff Optimal Strategy: (Everything on Heads) Degenerate Gambling (Everything on Tails)

15 Why not use value-of-information Capital Probability Rounds Rate of return (R)

16 Why not use value-of-information Rate of return: Ri Probability Capital at time i + == Capital at time i Long-run behavior: E[ R R2 R3 Rn ] Rate of return (R)

17 Why not use value-of-information Rate of return: Ri Probability Capital at time i + == Capital at time i Long-run behavior: E[ R R2 R3 Rn ] Converges to 0 in probability as n Rate of return (R)

18 Optimal reinvestment Daniel Bernoulli ( ) John Larry Kelly, Jr. ( )

19 Optimal reinvestment Doubling rate: Capital at time i + Wi == log Capital at time i (so R = 2W)

20 Optimal reinvestment Doubling rate: Capital at time i + Wi == log Capital at time i (so R = 2W) Long-run behavior: E[ R R2 R3 Rn ] == E[2W + W2 + W3 + + Wn] == 2E[W + W2 + W3 + + Wn] 2nE[W] for n

21 Optimal reinvestment Logarithmic expectation E[W] == p log bo is maximized by proportional gambling (b* == p). Arithmetic expectation E[R] == pbo is maximized by degenerate gambling

22 Measuring relevant information $! Amount of relevant information! === $$ $ $$ Posterior expected doubling rate Prior expected doubling rate

23 Measuring relevant information Definition (Relevant Information): For an agent with utility function u, the amount of relevant information contained in the message Y == y is K(y) == maxs Pr(x y) log u(s, x) Posterior optimal doubling rate maxs Pr(x) log u(s, x) Prior optimal doubling rate

24 Measuring relevant information K(y) == maxs Pr(x y) log u(s, x) maxs Pr(x) log u(s, x) Expected relevant information is non-negative. Relevant information equals the maximal fraction of future gains you can pay for a piece of information without loss. When u has the form u(s, x) == v(x) s(x) for some non-negative function v, relevant information equals Shannon information.

25 Example: Code-breaking

26 Example: Code-breaking Entropy: H=4 Accumulated information: I(X; Y) == 0

27 Example: Code-breaking bit! Entropy: H=3 Accumulated information: I(X; Y) ==

28 Example: Code-breaking 0 bit! Entropy: H=2 Accumulated information: I(X; Y) == 2

29 Example: Code-breaking 0 bit! Entropy: H= Accumulated information: I(X; Y) == 3

30 Example: Code-breaking 0 bit! Entropy: H=0 Accumulated information: I(X; Y) == 4

31 Example: Code-breaking 0 bit bit bit bit Entropy: H=0 Accumulated information: I(X; Y) == 4

32 Example: Code-breaking Rules: You can invest a fraction f of your capital in the guessing game If you guess the correct code, you get your investment back 6-fold: u == f + 6 f Otherwise, you lose it: u == f 5 W(f) == log( f ) + log( f + 6f ) 6 6

33 Example: Code-breaking Optimal strategy: f * == 0 Optimal doubling rate: W(f *) == W(f) == log( f ) + log( f + 6f ) 6 6

34 Example: Code-breaking 0.04 bits Optimal strategy: f * == /5 Optimal doubling rate: W(f *) == log( f ) + log( f + 6f ) W(f) == 8 8

35 Example: Code-breaking bits Optimal strategy: f * == 3/5 Optimal doubling rate: W(f *) == log( f ) + log( f + 6f ) W(f) == 4 4

36 Example: Code-breaking bits Optimal strategy: f * == 7/5 Optimal doubling rate: W(f *) ==.05 log( f ) + log( f + 6f ) W(f) == 2 2

37 Example: Code-breaking bits Optimal strategy: f * == Optimal doubling rate: W(f *) == log( f ) + log( f + 6f ) W(f) ==

38 Example: Code-breaking Raw information (drop in entropy) Relevant information (increase in doubling rate)

39 Example: Randomization

40 Example: Randomization def choose(): if flip(): if flip(): return ROCK /3, /3, /3 else: return PAPER else: return SCISSORS /2, /4, /4

41 Example: Randomization Rules: 2 You () and the adversary (2) both bet $ You move first The winner takes the whole pool W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }

42 Example: Randomization Best accessible strategy: p* == (, 0, 0) Doubling rate: W(p*) == W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }

43 Example: Randomization Best accessible strategy: p* == (/2, /2, 0) Doubling rate: W(p*) ==.00 W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }

44 Example: Randomization Best accessible strategy: p* == (2/4, /4, /4) Doubling rate: W(p*) == 0.42 W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }

45 Example: Randomization Best accessible strategy: p* == (3/8, 3/8, 2/8) Doubling rate: W(p*) == 0.9 W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }

46 Example: Randomization Best accessible strategy: p* == (6/6, 5/6, 5/6) Doubling rate: W(p*) == 0.09 W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }

47 Example: Randomization Coin flips Distribution Doubling rate 0 (, 0, 0) (/2, /2, 0) (/2, /4, /4) (3/8, 3/8, 2/8) (6/6, 5/6, 5/6) (/3, /3, /3)

48

49 January: Project course in information theory ith w w No E MOR N! O N SHAN Day : Uncertainty and Inference Probability theory: Semantics and expressivity Random variables Generative Bayesian models stochastic processes Uncertain and information: Uncertainty as cost The Hartley measure Shannon information content and entropy Huffman coding Day 2: Counting Typical Sequences Day 3: Guessing and Gambling Evidence, likelihood ratios, competitive prediction Kullback-Leibler divergence Examples of diverging stochastic models Expressivity and the bias/variance tradeoffs. Doubling rates and proportional betting Card color prediction Day 4: Asking Questions and Engineering Answers Questions and answers (or experiments and observations) mutual information Coin weighing The maximum entropy principle The channel coding theorem Day 5: Informative Descriptions and Residual Randomness The law of large numbers Typical sequences and the source coding theorem. The practical problem of source coding Kraft s inequality and prefix codes Arithmetic coding Stochastic processes and entropy rates the source coding theorem for stochastic processes Examples Kolmogorov complexity Tests of randomness Asymptotic equivalence of complexity and entropy

50

51

Information Theory and Communication

Information Theory and Communication Information Theory and Communication Shannon-Fano-Elias Code and Arithmetic Codes Ritwik Banerjee rbanerjee@cs.stonybrook.edu c Ritwik Banerjee Information Theory and Communication 1/12 Roadmap Examples

More information

CSE 203A: Randomized Algorithms

CSE 203A: Randomized Algorithms CSE 203A: Randomized Algorithms (00:30) The Minimax Principle (Ch 2) Lecture on 10/18/2017 by Daniel Kane Notes by Peter Greer-Berezovsky Game Tree Evaluation (2.1) - n round deterministic 2-player game

More information

Compressing Data. Konstantin Tretyakov

Compressing Data. Konstantin Tretyakov Compressing Data Konstantin Tretyakov (kt@ut.ee) MTAT.03.238 Advanced April 26, 2012 Claude Elwood Shannon (1916-2001) C. E. Shannon. A mathematical theory of communication. 1948 C. E. Shannon. The mathematical

More information

Security Models: Proofs, Protocols and Certification

Security Models: Proofs, Protocols and Certification Security Models: Proofs, Protocols and Certification Florent Autrau - Yassine Lakhnech - Jean-Louis Roch Master-2 Security, Cryptology and Coding of Information Systems ENSIMAG/Grenoble-INP UJF Grenoble

More information

Modeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA

Modeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA Modeling and Reasoning with Bayesian Networks Adnan Darwiche University of California Los Angeles, CA darwiche@cs.ucla.edu June 24, 2008 Contents Preface 1 1 Introduction 1 1.1 Automated Reasoning........................

More information

Computer vision: models, learning and inference. Chapter 10 Graphical Models

Computer vision: models, learning and inference. Chapter 10 Graphical Models Computer vision: models, learning and inference Chapter 10 Graphical Models Independence Two variables x 1 and x 2 are independent if their joint probability distribution factorizes as Pr(x 1, x 2 )=Pr(x

More information

Multiple Agents. Why can t we all just get along? (Rodney King) CS 3793/5233 Artificial Intelligence Multiple Agents 1

Multiple Agents. Why can t we all just get along? (Rodney King) CS 3793/5233 Artificial Intelligence Multiple Agents 1 Multiple Agents Why can t we all just get along? (Rodney King) CS 3793/5233 Artificial Intelligence Multiple Agents 1 Assumptions Assumptions Definitions Partially bservable Each agent can act autonomously.

More information

COMPSCI 650 Applied Information Theory Feb 2, Lecture 5. Recall the example of Huffman Coding on a binary string from last class:

COMPSCI 650 Applied Information Theory Feb 2, Lecture 5. Recall the example of Huffman Coding on a binary string from last class: COMPSCI 650 Applied Information Theory Feb, 016 Lecture 5 Instructor: Arya Mazumdar Scribe: Larkin Flodin, John Lalor 1 Huffman Coding 1.1 Last Class s Example Recall the example of Huffman Coding on a

More information

Is the result of multiplying 2 positive numbers positive or negative? Is the result of multiplying 2 negative numbers positive or negative?

Is the result of multiplying 2 positive numbers positive or negative? Is the result of multiplying 2 negative numbers positive or negative? Write down all the work necessary for the problem, this is a part of your class notes that you can use to study from. You will not be able to use notes on quizzes or tests. DO NOT simply write it and toss

More information

Graph Contraction. Graph Contraction CSE341T/CSE549T 10/20/2014. Lecture 14

Graph Contraction. Graph Contraction CSE341T/CSE549T 10/20/2014. Lecture 14 CSE341T/CSE549T 10/20/2014 Lecture 14 Graph Contraction Graph Contraction So far we have mostly talking about standard techniques for solving problems on graphs that were developed in the context of sequential

More information

Measuring Intrusion Detection Capability: An Information- Theoretic Approach

Measuring Intrusion Detection Capability: An Information- Theoretic Approach Measuring Intrusion Detection Capability: An Information- Theoretic Approach Guofei Gu, Prahlad Fogla, David Dagon, Wenke Lee Georgia Tech Boris Skoric Philips Research Lab Outline Motivation Problem Why

More information

Bayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19

Bayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19 Bayesian Methods David Rosenberg New York University April 11, 2017 David Rosenberg (New York University) DS-GA 1003 April 11, 2017 1 / 19 Classical Statistics Classical Statistics David Rosenberg (New

More information

Course: Algebra MP: Reason abstractively and quantitatively MP: Model with mathematics MP: Look for and make use of structure

Course: Algebra MP: Reason abstractively and quantitatively MP: Model with mathematics MP: Look for and make use of structure Modeling Cluster: Interpret the structure of expressions. A.SSE.1: Interpret expressions that represent a quantity in terms of its context. a. Interpret parts of an expression, such as terms, factors,

More information

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)

More information

The Cross-Entropy Method for Mathematical Programming

The Cross-Entropy Method for Mathematical Programming The Cross-Entropy Method for Mathematical Programming Dirk P. Kroese Reuven Y. Rubinstein Department of Mathematics, The University of Queensland, Australia Faculty of Industrial Engineering and Management,

More information

Data Structures and Algorithms

Data Structures and Algorithms Data Structures and Algorithms Session 26. April 29, 2009 Instructor: Bert Huang http://www.cs.columbia.edu/~bert/courses/3137 Announcements Homework 6 due before last class: May 4th Final Review May 4th

More information

Data Mining Classification: Bayesian Decision Theory

Data Mining Classification: Bayesian Decision Theory Data Mining Classification: Bayesian Decision Theory Lecture Notes for Chapter 2 R. O. Duda, P. E. Hart, and D. G. Stork, Pattern classification, 2nd ed. New York: Wiley, 2001. Lecture Notes for Chapter

More information

CS 188: Artificial Intelligence Spring Announcements

CS 188: Artificial Intelligence Spring Announcements CS 188: Artificial Intelligence Spring 2011 Lecture 20: Naïve Bayes 4/11/2011 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein. W4 due right now Announcements P4 out, due Friday First contest competition

More information

Autoencoders. Stephen Scott. Introduction. Basic Idea. Stacked AE. Denoising AE. Sparse AE. Contractive AE. Variational AE GAN.

Autoencoders. Stephen Scott. Introduction. Basic Idea. Stacked AE. Denoising AE. Sparse AE. Contractive AE. Variational AE GAN. Stacked Denoising Sparse Variational (Adapted from Paul Quint and Ian Goodfellow) Stacked Denoising Sparse Variational Autoencoding is training a network to replicate its input to its output Applications:

More information

Math 8 SOL Review

Math 8 SOL Review Math 8 SOL Review 2011-2012 SOL 8.1 The student will a) simplify numerical expressions involving positive exponents, using rational numbers, order of operations, and properties of operations with real

More information

Welcome to the course Algorithm Design

Welcome to the course Algorithm Design Welcome to the course Algorithm Design Summer Term 2011 Friedhelm Meyer auf der Heide Lecture 13, 15.7.2011 Friedhelm Meyer auf der Heide 1 Topics - Divide & conquer - Dynamic programming - Greedy Algorithms

More information

Problems 1 and 5 were graded by Amin Sorkhei, Problems 2 and 3 by Johannes Verwijnen and Problem 4 by Jyrki Kivinen. Entropy(D) = Gini(D) = 1

Problems 1 and 5 were graded by Amin Sorkhei, Problems 2 and 3 by Johannes Verwijnen and Problem 4 by Jyrki Kivinen. Entropy(D) = Gini(D) = 1 Problems and were graded by Amin Sorkhei, Problems and 3 by Johannes Verwijnen and Problem by Jyrki Kivinen.. [ points] (a) Gini index and Entropy are impurity measures which can be used in order to measure

More information

A HYBRID APPROACH FOR HANDLING UNCERTAINTY - PROBABILISTIC THEORY, CERTAINTY FACTOR AND FUZZY LOGIC

A HYBRID APPROACH FOR HANDLING UNCERTAINTY - PROBABILISTIC THEORY, CERTAINTY FACTOR AND FUZZY LOGIC Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 2, Issue. 11, November 2013,

More information

Multimedia Networking ECE 599

Multimedia Networking ECE 599 Multimedia Networking ECE 599 Prof. Thinh Nguyen School of Electrical Engineering and Computer Science Based on B. Lee s lecture notes. 1 Outline Compression basics Entropy and information theory basics

More information

Modelling image complexity by independent component analysis, with application to content-based image retrieval

Modelling image complexity by independent component analysis, with application to content-based image retrieval Modelling image complexity by independent component analysis, with application to content-based image retrieval Jukka Perkiö 1,2 and Aapo Hyvärinen 1,2,3 1 Helsinki Institute for Information Technology,

More information

Proving Lower bounds using Information Theory

Proving Lower bounds using Information Theory Proving Lower bounds using Information Theory Frédéric Magniez CNRS, Univ Paris Diderot Paris, July 2013 Data streams 2 Massive data sets - Examples: Network monitoring - Genome decoding Web databases

More information

Year 8 Mathematics Curriculum Map

Year 8 Mathematics Curriculum Map Year 8 Mathematics Curriculum Map Topic Algebra 1 & 2 Number 1 Title (Levels of Exercise) Objectives Sequences *To generate sequences using term-to-term and position-to-term rule. (5-6) Quadratic Sequences

More information

Virtual Vector Machine for Bayesian Online Classification

Virtual Vector Machine for Bayesian Online Classification Virtual Vector Machine for Bayesian Online Classification Thomas P. Minka, Rongjing Xiang, Yuan (Alan) Qi Appeared in UAI 2009 Presented by Lingbo Li Introduction Online Learning Update model parameters

More information

CS 188: Artificial Intelligence Fall Machine Learning

CS 188: Artificial Intelligence Fall Machine Learning CS 188: Artificial Intelligence Fall 2007 Lecture 23: Naïve Bayes 11/15/2007 Dan Klein UC Berkeley Machine Learning Up till now: how to reason or make decisions using a model Machine learning: how to select

More information

CONDITION CONTROLLED LOOPS. Introduction to Programming - Python

CONDITION CONTROLLED LOOPS. Introduction to Programming - Python CONDITION CONTROLLED LOOPS Introduction to Programming - Python Generating Random Numbers Generating a random integer Sometimes you need your program to generate information that isn t available when you

More information

Using Machine Learning to Optimize Storage Systems

Using Machine Learning to Optimize Storage Systems Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation

More information

VoI for adversarial information structures

VoI for adversarial information structures ARO ARO MURI MURI on on Value-centered Theory for for Adaptive Learning, Inference, Tracking, and and Exploitation VoI for adversarial information structures Kickf Emre Ertin, Nithin Sugavanam The Ohio

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 2: Probability: Discrete Random Variables Classification: Validation & Model Selection Many figures

More information

CS242: Probabilistic Graphical Models Lecture 3: Factor Graphs & Variable Elimination

CS242: Probabilistic Graphical Models Lecture 3: Factor Graphs & Variable Elimination CS242: Probabilistic Graphical Models Lecture 3: Factor Graphs & Variable Elimination Instructor: Erik Sudderth Brown University Computer Science September 11, 2014 Some figures and materials courtesy

More information

Probabilistic Learning Classification using Naïve Bayes

Probabilistic Learning Classification using Naïve Bayes Probabilistic Learning Classification using Naïve Bayes Weather forecasts are usually provided in terms such as 70 percent chance of rain. These forecasts are known as probabilities of precipitation reports.

More information

Wireless Network Security : Spring Arjun Athreya March 3, 2011 Survey: Trust Evaluation

Wireless Network Security : Spring Arjun Athreya March 3, 2011 Survey: Trust Evaluation Wireless Network Security 18-639: Spring 2011 Arjun Athreya March 3, 2011 Survey: Trust Evaluation A scenario LOBOS Management Co A CMU grad student new to Pittsburgh is looking for housing options in

More information

Detection and Mitigation of Cyber-Attacks using Game Theory

Detection and Mitigation of Cyber-Attacks using Game Theory Detection and Mitigation of Cyber-Attacks using Game Theory João P. Hespanha Kyriakos G. Vamvoudakis Correlation Engine COAs Data Data Data Data Cyber Situation Awareness Framework Mission Cyber-Assets

More information

CS 188: Artificial Intelligence Fall Announcements

CS 188: Artificial Intelligence Fall Announcements CS 188: Artificial Intelligence Fall 2006 Lecture 22: Naïve Bayes 11/14/2006 Dan Klein UC Berkeley Announcements Optional midterm On Tuesday 11/21 in class Review session 11/19, 7-9pm, in 306 Soda Projects

More information

Announcements. CS 188: Artificial Intelligence Fall Machine Learning. Classification. Classification. Bayes Nets for Classification

Announcements. CS 188: Artificial Intelligence Fall Machine Learning. Classification. Classification. Bayes Nets for Classification CS 88: Artificial Intelligence Fall 00 Lecture : Naïve Bayes //00 Announcements Optional midterm On Tuesday / in class Review session /9, 7-9pm, in 0 Soda Projects. due /. due /7 Dan Klein UC Berkeley

More information

Decentralized Stochastic Planning for Nonparametric Bayesian Models

Decentralized Stochastic Planning for Nonparametric Bayesian Models Decentralized Stochastic Planning for Nonparametric Bayesian Models Silvia Ferrari Professor of Engineering and Computer Science Department of Mechanical Engineering and Materials Science Duke University

More information

Chapter 5: Data compression. Chapter 5 outline

Chapter 5: Data compression. Chapter 5 outline Chapter 5: Data compression Chapter 5 outline 2 balls weighing problem Examples of codes Kraft inequality Optimal codes + bounds Kraft inequality for uniquely decodable codes Huffman codes Shannon-Fano-Elias

More information

The PageRank Computation in Google, Randomized Algorithms and Consensus of Multi-Agent Systems

The PageRank Computation in Google, Randomized Algorithms and Consensus of Multi-Agent Systems The PageRank Computation in Google, Randomized Algorithms and Consensus of Multi-Agent Systems Roberto Tempo IEIIT-CNR Politecnico di Torino tempo@polito.it This talk The objective of this talk is to discuss

More information

Meshkat Farrokhzadi. in Partial Fulfillment of the Requirements for the Degree of. May 25, Certified b rvisor

Meshkat Farrokhzadi. in Partial Fulfillment of the Requirements for the Degree of. May 25, Certified b rvisor Entropy, Information Rate and Mutual Information Measures for the Email Content of Information Workers by Meshkat Farrokhzadi Submitted to the Department of Electrical Engineering and Computer Science

More information

SAR change detection based on Generalized Gamma distribution. divergence and auto-threshold segmentation

SAR change detection based on Generalized Gamma distribution. divergence and auto-threshold segmentation SAR change detection based on Generalized Gamma distribution divergence and auto-threshold segmentation GAO Cong-shan 1 2, ZHANG Hong 1*, WANG Chao 1 1.Center for Earth Observation and Digital Earth, CAS,

More information

Bayesian model selection and diagnostics

Bayesian model selection and diagnostics Bayesian model selection and diagnostics A typical Bayesian analysis compares a handful of models. Example 1: Consider the spline model for the motorcycle data, how many basis functions? Example 2: Consider

More information

Lecture 1: An Introduction to Online Algorithms

Lecture 1: An Introduction to Online Algorithms Algoritmos e Incerteza (PUC-Rio INF979, 017.1) Lecture 1: An Introduction to Online Algorithms Mar 1, 017 Lecturer: Marco Molinaro Scribe: Joao Pedro T. Brandao Online algorithms differ from traditional

More information

Lexical Semantics. Regina Barzilay MIT. October, 5766

Lexical Semantics. Regina Barzilay MIT. October, 5766 Lexical Semantics Regina Barzilay MIT October, 5766 Last Time: Vector-Based Similarity Measures man woman grape orange apple n Euclidian: x, y = x y = i=1 ( x i y i ) 2 n x y x i y i i=1 Cosine: cos( x,

More information

AI Programming CS S-15 Probability Theory

AI Programming CS S-15 Probability Theory AI Programming CS662-2013S-15 Probability Theory David Galles Department of Computer Science University of San Francisco 15-0: Uncertainty In many interesting agent environments, uncertainty plays a central

More information

Contents. 3 Vector Quantization The VQ Advantage Formulation Optimality Conditions... 48

Contents. 3 Vector Quantization The VQ Advantage Formulation Optimality Conditions... 48 Contents Part I Prelude 1 Introduction... 3 1.1 Audio Coding... 4 1.2 Basic Idea... 6 1.3 Perceptual Irrelevance... 8 1.4 Statistical Redundancy... 9 1.5 Data Modeling... 9 1.6 Resolution Challenge...

More information

CS154, Lecture 18: 1

CS154, Lecture 18: 1 CS154, Lecture 18: 1 CS 154 Final Exam Wednesday December 13, 12:15-3:15 pm Skilling Auditorium You re allowed one double-sided sheet of notes Exam is comprehensive (but will emphasize post-midterm topics)

More information

round decimals to the nearest decimal place and order negative numbers in context

round decimals to the nearest decimal place and order negative numbers in context 6 Numbers and the number system understand and use proportionality use the equivalence of fractions, decimals and percentages to compare proportions use understanding of place value to multiply and divide

More information

Lecture 17. Lower bound for variable-length source codes with error. Coding a sequence of symbols: Rates and scheme (Arithmetic code)

Lecture 17. Lower bound for variable-length source codes with error. Coding a sequence of symbols: Rates and scheme (Arithmetic code) Lecture 17 Agenda for the lecture Lower bound for variable-length source codes with error Coding a sequence of symbols: Rates and scheme (Arithmetic code) Introduction to universal codes 17.1 variable-length

More information

Crypto-systems all around us ATM machines Remote logins using SSH Web browsers (https invokes Secure Socket Layer (SSL))

Crypto-systems all around us ATM machines Remote logins using SSH Web browsers (https invokes Secure Socket Layer (SSL)) Introduction (Mihir Bellare Text/Notes: http://cseweb.ucsd.edu/users/mihir/cse207/) Cryptography provides: Data Privacy Data Integrity and Authenticity Crypto-systems all around us ATM machines Remote

More information

FOUNDATION FOR DATA RESOURCE QUALITY

FOUNDATION FOR DATA RESOURCE QUALITY FOUNDATION FOR DATA RESOURCE QUALITY CENTRAL OHIO DAMA APRIL 14, 2005 MICHAEL H. BRACKETT CONSULTING DATA ARCHITECT DATA RESOURCE DESIGN & REMODELING DRDR P.O. Box 151 Lilliwaup, Washington 98555 Voice:

More information

Name. Final Exam, Economics 210A, December 2012 There are 8 questions. Answer as many as you can... Good luck!

Name. Final Exam, Economics 210A, December 2012 There are 8 questions. Answer as many as you can... Good luck! Name Final Exam, Economics 210A, December 2012 There are 8 questions. Answer as many as you can... Good luck! 1) Let S and T be convex sets in Euclidean n space. Let S + T be the set {x x = s + t for some

More information

Deep Learning Cook Book

Deep Learning Cook Book Deep Learning Cook Book Robert Haschke (CITEC) Overview Input Representation Output Layer + Cost Function Hidden Layer Units Initialization Regularization Input representation Choose an input representation

More information

Bayesian Classification Using Probabilistic Graphical Models

Bayesian Classification Using Probabilistic Graphical Models San Jose State University SJSU ScholarWorks Master's Projects Master's Theses and Graduate Research Spring 2014 Bayesian Classification Using Probabilistic Graphical Models Mehal Patel San Jose State University

More information

Adversary Arguments. A method for obtaining lower bounds

Adversary Arguments. A method for obtaining lower bounds Adversary Arguments A method for obtaining lower bounds What is an Adversary? A Second Algorithm Which Intercepts Access to Data Structures Constructs the input data only as needed Attempts to make original

More information

The PageRank Computation in Google, Randomized Algorithms and Consensus of Multi-Agent Systems

The PageRank Computation in Google, Randomized Algorithms and Consensus of Multi-Agent Systems The PageRank Computation in Google, Randomized Algorithms and Consensus of Multi-Agent Systems Roberto Tempo IEIIT-CNR Politecnico di Torino tempo@polito.it This talk The objective of this talk is to discuss

More information

Lecture 15. Error-free variable length schemes: Shannon-Fano code

Lecture 15. Error-free variable length schemes: Shannon-Fano code Lecture 15 Agenda for the lecture Bounds for L(X) Error-free variable length schemes: Shannon-Fano code 15.1 Optimal length nonsingular code While we do not know L(X), it is easy to specify a nonsingular

More information

Thresholds Determination for Probabilistic Rough Sets with Genetic Algorithms

Thresholds Determination for Probabilistic Rough Sets with Genetic Algorithms Thresholds Determination for Probabilistic Rough Sets with Genetic Algorithms Babar Majeed, Nouman Azam, JingTao Yao Department of Computer Science University of Regina {majeed2b,azam200n,jtyao}@cs.uregina.ca

More information

Alternatives to Direct Supervision

Alternatives to Direct Supervision CreativeAI: Deep Learning for Graphics Alternatives to Direct Supervision Niloy Mitra Iasonas Kokkinos Paul Guerrero Nils Thuerey Tobias Ritschel UCL UCL UCL TUM UCL Timetable Theory and Basics State of

More information

Gradient Descent. Wed Sept 20th, James McInenrey Adapted from slides by Francisco J. R. Ruiz

Gradient Descent. Wed Sept 20th, James McInenrey Adapted from slides by Francisco J. R. Ruiz Gradient Descent Wed Sept 20th, 2017 James McInenrey Adapted from slides by Francisco J. R. Ruiz Housekeeping A few clarifications of and adjustments to the course schedule: No more breaks at the midpoint

More information

Extreme Value Theory in (Hourly) Precipitation

Extreme Value Theory in (Hourly) Precipitation Extreme Value Theory in (Hourly) Precipitation Uli Schneider Geophysical Statistics Project, NCAR GSP Miniseries at CSU November 17, 2003 Outline Project overview Extreme value theory 101 Applying extreme

More information

CAP 5993/CAP 4993 Game Theory. Instructor: Sam Ganzfried

CAP 5993/CAP 4993 Game Theory. Instructor: Sam Ganzfried CAP 5993/CAP 4993 Game Theory Instructor: Sam Ganzfried sganzfri@cis.fiu.edu 1 Announcements HW 1 due today HW 2 out this week (2/2), due 2/14 2 Definition: A two-player game is a zero-sum game if for

More information

Computational Cognitive Science

Computational Cognitive Science Computational Cognitive Science Lecture 5: Maximum Likelihood Estimation; Parameter Uncertainty Chris Lucas (Slides adapted from Frank Keller s) School of Informatics University of Edinburgh clucas2@inf.ed.ac.uk

More information

LAB 5: SELECTION STATEMENTS

LAB 5: SELECTION STATEMENTS Statement Purpose: The purpose of this lab is to familiarize students with Conditional statements and explain how to control the sequence of statement execution, depending on the value of an expression

More information

INLA: Integrated Nested Laplace Approximations

INLA: Integrated Nested Laplace Approximations INLA: Integrated Nested Laplace Approximations John Paige Statistics Department University of Washington October 10, 2017 1 The problem Markov Chain Monte Carlo (MCMC) takes too long in many settings.

More information

1.1 calculator viewing window find roots in your calculator 1.2 functions find domain and range (from a graph) may need to review interval notation

1.1 calculator viewing window find roots in your calculator 1.2 functions find domain and range (from a graph) may need to review interval notation 1.1 calculator viewing window find roots in your calculator 1.2 functions find domain and range (from a graph) may need to review interval notation functions vertical line test function notation evaluate

More information

Communication Amid Uncertainty

Communication Amid Uncertainty Communication Amid Uncertainty Madhu Sudan Microsoft Research Based on Juba, S. (STOC 2008, ITCS 2011) Goldreich, Juba, S. (JACM 2011) Juba, Kalai, Khanna, S. (ITCS 2011) Haramaty, S. (ITCS 2014) Canonne,

More information

Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay

Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 29 Source Coding (Part-4) We have already had 3 classes on source coding

More information

Exam Topics. Search in Discrete State Spaces. What is intelligence? Adversarial Search. Which Algorithm? 6/1/2012

Exam Topics. Search in Discrete State Spaces. What is intelligence? Adversarial Search. Which Algorithm? 6/1/2012 Exam Topics Artificial Intelligence Recap & Expectation Maximization CSE 473 Dan Weld BFS, DFS, UCS, A* (tree and graph) Completeness and Optimality Heuristics: admissibility and consistency CSPs Constraint

More information

Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay

Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 26 Source Coding (Part 1) Hello everyone, we will start a new module today

More information

A Well-Behaved Algorithm for Simulating Dependence Structures of Bayesian Networks

A Well-Behaved Algorithm for Simulating Dependence Structures of Bayesian Networks A Well-Behaved Algorithm for Simulating Dependence Structures of Bayesian Networks Yang Xiang and Tristan Miller Department of Computer Science University of Regina Regina, Saskatchewan, Canada S4S 0A2

More information

Clustering Using Elements of Information Theory

Clustering Using Elements of Information Theory Clustering Using Elements of Information Theory Daniel de Araújo 1,2, Adrião Dória Neto 2, Jorge Melo 2, and Allan Martins 2 1 Federal Rural University of Semi-Árido, Campus Angicos, Angicos/RN, Brasil

More information

Communication Amid Uncertainty

Communication Amid Uncertainty Communication Amid Uncertainty Madhu Sudan Microsoft Research Based on Juba, S. (STOC 2008, ITCS 2011) Goldreich, Juba, S. (JACM 2011) Juba, Kalai, Khanna, S. (ITCS 2011) Haramaty, S. (ITCS 2014) Canonne,

More information

What is a Computer Algorithm?

What is a Computer Algorithm? What is a Computer Algorithm? A computer algorithm is a detailed step-by-step method for solving a problem by using a computer. Problem-Solving (Science and Engineering) Analysis How does it work? Breaking

More information

Computer security experts often tell us not to choose an easy to guess word as a password. For example, the user name or date of birth.

Computer security experts often tell us not to choose an easy to guess word as a password. For example, the user name or date of birth. Graphical Passwords: Computer security experts often tell us not to choose an easy to guess word as a password. For example, the user name or date of birth. The Graphical Passwords team have been working

More information

EE67I Multimedia Communication Systems Lecture 4

EE67I Multimedia Communication Systems Lecture 4 EE67I Multimedia Communication Systems Lecture 4 Lossless Compression Basics of Information Theory Compression is either lossless, in which no information is lost, or lossy in which information is lost.

More information

Randomized algorithms have several advantages over deterministic ones. We discuss them here:

Randomized algorithms have several advantages over deterministic ones. We discuss them here: CS787: Advanced Algorithms Lecture 6: Randomized Algorithms In this lecture we introduce randomized algorithms. We will begin by motivating the use of randomized algorithms through a few examples. Then

More information

David Rappaport School of Computing Queen s University CANADA. Copyright, 1996 Dale Carnegie & Associates, Inc.

David Rappaport School of Computing Queen s University CANADA. Copyright, 1996 Dale Carnegie & Associates, Inc. David Rappaport School of Computing Queen s University CANADA Copyright, 1996 Dale Carnegie & Associates, Inc. Data Compression There are two broad categories of data compression: Lossless Compression

More information

Copyright 2000, Kevin Wayne 1

Copyright 2000, Kevin Wayne 1 Guessing Game: NP-Complete? 1. LONGEST-PATH: Given a graph G = (V, E), does there exists a simple path of length at least k edges? YES. SHORTEST-PATH: Given a graph G = (V, E), does there exists a simple

More information

Montana City School GRADE 5

Montana City School GRADE 5 Montana City School GRADE 5 Montana Standard 1: Students engage in the mathematical processes of problem solving and reasoning, estimation, communication, connections and applications, and using appropriate

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Other models of interactive domains Marc Toussaint University of Stuttgart Winter 2018/19 Basic Taxonomy of domain models Other models of interactive domains Basic Taxonomy of domain

More information

Welcome to the course Algorithm Design

Welcome to the course Algorithm Design Welcome to the course Algorithm Design Summer Term 2011 Friedhelm Meyer auf der Heide Lecture 12, 8.7.2011 Friedhelm Meyer auf der Heide 1 Randomised Algorithms Friedhelm Meyer auf der Heide 2 Topics -

More information

List of NEW Maths content

List of NEW Maths content List of NEW Maths content Our brand new Maths content for the new Maths GCSE (9-1) consists of 212 chapters broken up into 37 titles and 4 topic areas (Algebra, Geometry & Measures, Number and Statistics).

More information

Performance Level Descriptors. Mathematics

Performance Level Descriptors. Mathematics Performance Level Descriptors Grade 3 Well Students rarely, Understand that our number system is based on combinations of 1s, 10s, and 100s (place value, compare, order, decompose, and combine using addition)

More information

Final Exam. Introduction to Artificial Intelligence. CS 188 Spring 2010 INSTRUCTIONS. You have 3 hours.

Final Exam. Introduction to Artificial Intelligence. CS 188 Spring 2010 INSTRUCTIONS. You have 3 hours. CS 188 Spring 2010 Introduction to Artificial Intelligence Final Exam INSTRUCTIONS You have 3 hours. The exam is closed book, closed notes except a two-page crib sheet. Please use non-programmable calculators

More information

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao Introduction to GAN Generative Adversarial Networks Junheng(Jeff) Hao Adversarial Training is the coolest thing since sliced bread. -- Yann LeCun Roadmap 1. Generative Modeling 2. GAN 101: What is GAN?

More information

Supervised vs unsupervised clustering

Supervised vs unsupervised clustering Classification Supervised vs unsupervised clustering Cluster analysis: Classes are not known a- priori. Classification: Classes are defined a-priori Sometimes called supervised clustering Extract useful

More information

Expectation-Maximization Methods in Population Analysis. Robert J. Bauer, Ph.D. ICON plc.

Expectation-Maximization Methods in Population Analysis. Robert J. Bauer, Ph.D. ICON plc. Expectation-Maximization Methods in Population Analysis Robert J. Bauer, Ph.D. ICON plc. 1 Objective The objective of this tutorial is to briefly describe the statistical basis of Expectation-Maximization

More information

Computer Vision. Exercise Session 10 Image Categorization

Computer Vision. Exercise Session 10 Image Categorization Computer Vision Exercise Session 10 Image Categorization Object Categorization Task Description Given a small number of training images of a category, recognize a-priori unknown instances of that category

More information

CDG2A/CDZ4A/CDC4A/ MBT4A ELEMENTS OF OPERATIONS RESEARCH. Unit : I - V

CDG2A/CDZ4A/CDC4A/ MBT4A ELEMENTS OF OPERATIONS RESEARCH. Unit : I - V CDG2A/CDZ4A/CDC4A/ MBT4A ELEMENTS OF OPERATIONS RESEARCH Unit : I - V UNIT I Introduction Operations Research Meaning and definition. Origin and History Characteristics and Scope Techniques in Operations

More information

Chapter 1 Polynomials and Modeling

Chapter 1 Polynomials and Modeling Chapter 1 Polynomials and Modeling 1.1 Linear Functions Recall that a line is a function of the form y = mx+ b, where m is the slope of the line (how steep the line is) and b gives the y-intercept (where

More information

CSE 421 Greedy: Huffman Codes

CSE 421 Greedy: Huffman Codes CSE 421 Greedy: Huffman Codes Yin Tat Lee 1 Compression Example 100k file, 6 letter alphabet: File Size: ASCII, 8 bits/char: 800kbits 2 3 > 6; 3 bits/char: 300kbits a 45% b 13% c 12% d 16% e 9% f 5% Why?

More information

Minnesota Academic Standards for Mathematics 2007

Minnesota Academic Standards for Mathematics 2007 An Alignment of Minnesota for Mathematics 2007 to the Pearson Integrated High School Mathematics 2014 to Pearson Integrated High School Mathematics Common Core Table of Contents Chapter 1... 1 Chapter

More information

Monte Carlo Methods and Statistical Computing: My Personal E

Monte Carlo Methods and Statistical Computing: My Personal E Monte Carlo Methods and Statistical Computing: My Personal Experience Department of Mathematics & Statistics Indian Institute of Technology Kanpur November 29, 2014 Outline Preface 1 Preface 2 3 4 5 6

More information

Math Interim Mini-Tests. 3rd Grade Mini-Tests

Math Interim Mini-Tests. 3rd Grade Mini-Tests 3rd Grade Mini-Tests Mini-Test Name Availability Area of Plane Figures-01 Gr 3_Model, Reason, & Solve Problems-04 Multiplicative Properties & Factors-01 Patterns & Using the Four Operations-01 Real-World

More information

Unsupervised Learning

Unsupervised Learning Deep Learning for Graphics Unsupervised Learning Niloy Mitra Iasonas Kokkinos Paul Guerrero Vladimir Kim Kostas Rematas Tobias Ritschel UCL UCL/Facebook UCL Adobe Research U Washington UCL Timetable Niloy

More information

Algorithms for Markov Random Fields in Computer Vision

Algorithms for Markov Random Fields in Computer Vision Algorithms for Markov Random Fields in Computer Vision Dan Huttenlocher November, 2003 (Joint work with Pedro Felzenszwalb) Random Field Broadly applicable stochastic model Collection of n sites S Hidden

More information