A Quantitative Measure of Relevance Based on Kelly Gambling Theory
|
|
- Gillian Caldwell
- 6 years ago
- Views:
Transcription
1 A Quantitative Measure of Relevance Based on Kelly Gambling Theory Mathias Winther Madsen Institute for Logic, Language, and Computation University of Amsterdam
2 PLAN Why How Examples
3 Why
4 Why
5 How
6 Why not use Shannon information H(X) == E log Pr(X == x) Claude Shannon (96 200)
7 Why not use Shannon information Information Content === Prior Uncertainty (cf. Klir 2008; Shannon 948) Posterior Uncertainty
8 Why not use Shannon information Pr(X == ) == 0.5 Pr(X == 2) == 0.9 What is the value of X Pr(X == 3) == 0.23 Pr(X == 4) == 0.2 H(X) == E log Pr(X == x) Pr(X == 5) == 0.22 == 2.3
9 Why not use Shannon information 0 Pr(X == ) == 0.5 Pr(X == 2) == 0.9 Is X == 2 0 Is X == 3 0 Pr(X == 3) == 0.23 Is X in {4,5} Is X == 5 Expected number of questions: == 2.34 Pr(X == 4) == Pr(X == 5) == 0.22
10 What color are my socks H(p) == p log p == 6.53 bits of entropy.
11 How
12 Why not use value-of-information $! Value-ofInformation! === $$ $ $$ Posterior Expectation Prior Expectation
13 Why not use value-of-information Rules: Your capital can be distributed freely Bets on the actual outcome are returned twofold Bets on all other outcomes are lost
14 Why not use value-of-information Expected payoff Optimal Strategy: (Everything on Heads) Degenerate Gambling (Everything on Tails)
15 Why not use value-of-information Capital Probability Rounds Rate of return (R)
16 Why not use value-of-information Rate of return: Ri Probability Capital at time i + == Capital at time i Long-run behavior: E[ R R2 R3 Rn ] Rate of return (R)
17 Why not use value-of-information Rate of return: Ri Probability Capital at time i + == Capital at time i Long-run behavior: E[ R R2 R3 Rn ] Converges to 0 in probability as n Rate of return (R)
18 Optimal reinvestment Daniel Bernoulli ( ) John Larry Kelly, Jr. ( )
19 Optimal reinvestment Doubling rate: Capital at time i + Wi == log Capital at time i (so R = 2W)
20 Optimal reinvestment Doubling rate: Capital at time i + Wi == log Capital at time i (so R = 2W) Long-run behavior: E[ R R2 R3 Rn ] == E[2W + W2 + W3 + + Wn] == 2E[W + W2 + W3 + + Wn] 2nE[W] for n
21 Optimal reinvestment Logarithmic expectation E[W] == p log bo is maximized by proportional gambling (b* == p). Arithmetic expectation E[R] == pbo is maximized by degenerate gambling
22 Measuring relevant information $! Amount of relevant information! === $$ $ $$ Posterior expected doubling rate Prior expected doubling rate
23 Measuring relevant information Definition (Relevant Information): For an agent with utility function u, the amount of relevant information contained in the message Y == y is K(y) == maxs Pr(x y) log u(s, x) Posterior optimal doubling rate maxs Pr(x) log u(s, x) Prior optimal doubling rate
24 Measuring relevant information K(y) == maxs Pr(x y) log u(s, x) maxs Pr(x) log u(s, x) Expected relevant information is non-negative. Relevant information equals the maximal fraction of future gains you can pay for a piece of information without loss. When u has the form u(s, x) == v(x) s(x) for some non-negative function v, relevant information equals Shannon information.
25 Example: Code-breaking
26 Example: Code-breaking Entropy: H=4 Accumulated information: I(X; Y) == 0
27 Example: Code-breaking bit! Entropy: H=3 Accumulated information: I(X; Y) ==
28 Example: Code-breaking 0 bit! Entropy: H=2 Accumulated information: I(X; Y) == 2
29 Example: Code-breaking 0 bit! Entropy: H= Accumulated information: I(X; Y) == 3
30 Example: Code-breaking 0 bit! Entropy: H=0 Accumulated information: I(X; Y) == 4
31 Example: Code-breaking 0 bit bit bit bit Entropy: H=0 Accumulated information: I(X; Y) == 4
32 Example: Code-breaking Rules: You can invest a fraction f of your capital in the guessing game If you guess the correct code, you get your investment back 6-fold: u == f + 6 f Otherwise, you lose it: u == f 5 W(f) == log( f ) + log( f + 6f ) 6 6
33 Example: Code-breaking Optimal strategy: f * == 0 Optimal doubling rate: W(f *) == W(f) == log( f ) + log( f + 6f ) 6 6
34 Example: Code-breaking 0.04 bits Optimal strategy: f * == /5 Optimal doubling rate: W(f *) == log( f ) + log( f + 6f ) W(f) == 8 8
35 Example: Code-breaking bits Optimal strategy: f * == 3/5 Optimal doubling rate: W(f *) == log( f ) + log( f + 6f ) W(f) == 4 4
36 Example: Code-breaking bits Optimal strategy: f * == 7/5 Optimal doubling rate: W(f *) ==.05 log( f ) + log( f + 6f ) W(f) == 2 2
37 Example: Code-breaking bits Optimal strategy: f * == Optimal doubling rate: W(f *) == log( f ) + log( f + 6f ) W(f) ==
38 Example: Code-breaking Raw information (drop in entropy) Relevant information (increase in doubling rate)
39 Example: Randomization
40 Example: Randomization def choose(): if flip(): if flip(): return ROCK /3, /3, /3 else: return PAPER else: return SCISSORS /2, /4, /4
41 Example: Randomization Rules: 2 You () and the adversary (2) both bet $ You move first The winner takes the whole pool W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }
42 Example: Randomization Best accessible strategy: p* == (, 0, 0) Doubling rate: W(p*) == W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }
43 Example: Randomization Best accessible strategy: p* == (/2, /2, 0) Doubling rate: W(p*) ==.00 W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }
44 Example: Randomization Best accessible strategy: p* == (2/4, /4, /4) Doubling rate: W(p*) == 0.42 W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }
45 Example: Randomization Best accessible strategy: p* == (3/8, 3/8, 2/8) Doubling rate: W(p*) == 0.9 W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }
46 Example: Randomization Best accessible strategy: p* == (6/6, 5/6, 5/6) Doubling rate: W(p*) == 0.09 W(p) == log min { p + 2 p2, p2 + 2 p3, p3 + 2 p }
47 Example: Randomization Coin flips Distribution Doubling rate 0 (, 0, 0) (/2, /2, 0) (/2, /4, /4) (3/8, 3/8, 2/8) (6/6, 5/6, 5/6) (/3, /3, /3)
48
49 January: Project course in information theory ith w w No E MOR N! O N SHAN Day : Uncertainty and Inference Probability theory: Semantics and expressivity Random variables Generative Bayesian models stochastic processes Uncertain and information: Uncertainty as cost The Hartley measure Shannon information content and entropy Huffman coding Day 2: Counting Typical Sequences Day 3: Guessing and Gambling Evidence, likelihood ratios, competitive prediction Kullback-Leibler divergence Examples of diverging stochastic models Expressivity and the bias/variance tradeoffs. Doubling rates and proportional betting Card color prediction Day 4: Asking Questions and Engineering Answers Questions and answers (or experiments and observations) mutual information Coin weighing The maximum entropy principle The channel coding theorem Day 5: Informative Descriptions and Residual Randomness The law of large numbers Typical sequences and the source coding theorem. The practical problem of source coding Kraft s inequality and prefix codes Arithmetic coding Stochastic processes and entropy rates the source coding theorem for stochastic processes Examples Kolmogorov complexity Tests of randomness Asymptotic equivalence of complexity and entropy
50
51
Information Theory and Communication
Information Theory and Communication Shannon-Fano-Elias Code and Arithmetic Codes Ritwik Banerjee rbanerjee@cs.stonybrook.edu c Ritwik Banerjee Information Theory and Communication 1/12 Roadmap Examples
More informationCSE 203A: Randomized Algorithms
CSE 203A: Randomized Algorithms (00:30) The Minimax Principle (Ch 2) Lecture on 10/18/2017 by Daniel Kane Notes by Peter Greer-Berezovsky Game Tree Evaluation (2.1) - n round deterministic 2-player game
More informationCompressing Data. Konstantin Tretyakov
Compressing Data Konstantin Tretyakov (kt@ut.ee) MTAT.03.238 Advanced April 26, 2012 Claude Elwood Shannon (1916-2001) C. E. Shannon. A mathematical theory of communication. 1948 C. E. Shannon. The mathematical
More informationSecurity Models: Proofs, Protocols and Certification
Security Models: Proofs, Protocols and Certification Florent Autrau - Yassine Lakhnech - Jean-Louis Roch Master-2 Security, Cryptology and Coding of Information Systems ENSIMAG/Grenoble-INP UJF Grenoble
More informationModeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA
Modeling and Reasoning with Bayesian Networks Adnan Darwiche University of California Los Angeles, CA darwiche@cs.ucla.edu June 24, 2008 Contents Preface 1 1 Introduction 1 1.1 Automated Reasoning........................
More informationComputer vision: models, learning and inference. Chapter 10 Graphical Models
Computer vision: models, learning and inference Chapter 10 Graphical Models Independence Two variables x 1 and x 2 are independent if their joint probability distribution factorizes as Pr(x 1, x 2 )=Pr(x
More informationMultiple Agents. Why can t we all just get along? (Rodney King) CS 3793/5233 Artificial Intelligence Multiple Agents 1
Multiple Agents Why can t we all just get along? (Rodney King) CS 3793/5233 Artificial Intelligence Multiple Agents 1 Assumptions Assumptions Definitions Partially bservable Each agent can act autonomously.
More informationCOMPSCI 650 Applied Information Theory Feb 2, Lecture 5. Recall the example of Huffman Coding on a binary string from last class:
COMPSCI 650 Applied Information Theory Feb, 016 Lecture 5 Instructor: Arya Mazumdar Scribe: Larkin Flodin, John Lalor 1 Huffman Coding 1.1 Last Class s Example Recall the example of Huffman Coding on a
More informationIs the result of multiplying 2 positive numbers positive or negative? Is the result of multiplying 2 negative numbers positive or negative?
Write down all the work necessary for the problem, this is a part of your class notes that you can use to study from. You will not be able to use notes on quizzes or tests. DO NOT simply write it and toss
More informationGraph Contraction. Graph Contraction CSE341T/CSE549T 10/20/2014. Lecture 14
CSE341T/CSE549T 10/20/2014 Lecture 14 Graph Contraction Graph Contraction So far we have mostly talking about standard techniques for solving problems on graphs that were developed in the context of sequential
More informationMeasuring Intrusion Detection Capability: An Information- Theoretic Approach
Measuring Intrusion Detection Capability: An Information- Theoretic Approach Guofei Gu, Prahlad Fogla, David Dagon, Wenke Lee Georgia Tech Boris Skoric Philips Research Lab Outline Motivation Problem Why
More informationBayesian Methods. David Rosenberg. April 11, New York University. David Rosenberg (New York University) DS-GA 1003 April 11, / 19
Bayesian Methods David Rosenberg New York University April 11, 2017 David Rosenberg (New York University) DS-GA 1003 April 11, 2017 1 / 19 Classical Statistics Classical Statistics David Rosenberg (New
More informationCourse: Algebra MP: Reason abstractively and quantitatively MP: Model with mathematics MP: Look for and make use of structure
Modeling Cluster: Interpret the structure of expressions. A.SSE.1: Interpret expressions that represent a quantity in terms of its context. a. Interpret parts of an expression, such as terms, factors,
More informationFMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu
FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)
More informationThe Cross-Entropy Method for Mathematical Programming
The Cross-Entropy Method for Mathematical Programming Dirk P. Kroese Reuven Y. Rubinstein Department of Mathematics, The University of Queensland, Australia Faculty of Industrial Engineering and Management,
More informationData Structures and Algorithms
Data Structures and Algorithms Session 26. April 29, 2009 Instructor: Bert Huang http://www.cs.columbia.edu/~bert/courses/3137 Announcements Homework 6 due before last class: May 4th Final Review May 4th
More informationData Mining Classification: Bayesian Decision Theory
Data Mining Classification: Bayesian Decision Theory Lecture Notes for Chapter 2 R. O. Duda, P. E. Hart, and D. G. Stork, Pattern classification, 2nd ed. New York: Wiley, 2001. Lecture Notes for Chapter
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 20: Naïve Bayes 4/11/2011 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein. W4 due right now Announcements P4 out, due Friday First contest competition
More informationAutoencoders. Stephen Scott. Introduction. Basic Idea. Stacked AE. Denoising AE. Sparse AE. Contractive AE. Variational AE GAN.
Stacked Denoising Sparse Variational (Adapted from Paul Quint and Ian Goodfellow) Stacked Denoising Sparse Variational Autoencoding is training a network to replicate its input to its output Applications:
More informationMath 8 SOL Review
Math 8 SOL Review 2011-2012 SOL 8.1 The student will a) simplify numerical expressions involving positive exponents, using rational numbers, order of operations, and properties of operations with real
More informationWelcome to the course Algorithm Design
Welcome to the course Algorithm Design Summer Term 2011 Friedhelm Meyer auf der Heide Lecture 13, 15.7.2011 Friedhelm Meyer auf der Heide 1 Topics - Divide & conquer - Dynamic programming - Greedy Algorithms
More informationProblems 1 and 5 were graded by Amin Sorkhei, Problems 2 and 3 by Johannes Verwijnen and Problem 4 by Jyrki Kivinen. Entropy(D) = Gini(D) = 1
Problems and were graded by Amin Sorkhei, Problems and 3 by Johannes Verwijnen and Problem by Jyrki Kivinen.. [ points] (a) Gini index and Entropy are impurity measures which can be used in order to measure
More informationA HYBRID APPROACH FOR HANDLING UNCERTAINTY - PROBABILISTIC THEORY, CERTAINTY FACTOR AND FUZZY LOGIC
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 2, Issue. 11, November 2013,
More informationMultimedia Networking ECE 599
Multimedia Networking ECE 599 Prof. Thinh Nguyen School of Electrical Engineering and Computer Science Based on B. Lee s lecture notes. 1 Outline Compression basics Entropy and information theory basics
More informationModelling image complexity by independent component analysis, with application to content-based image retrieval
Modelling image complexity by independent component analysis, with application to content-based image retrieval Jukka Perkiö 1,2 and Aapo Hyvärinen 1,2,3 1 Helsinki Institute for Information Technology,
More informationProving Lower bounds using Information Theory
Proving Lower bounds using Information Theory Frédéric Magniez CNRS, Univ Paris Diderot Paris, July 2013 Data streams 2 Massive data sets - Examples: Network monitoring - Genome decoding Web databases
More informationYear 8 Mathematics Curriculum Map
Year 8 Mathematics Curriculum Map Topic Algebra 1 & 2 Number 1 Title (Levels of Exercise) Objectives Sequences *To generate sequences using term-to-term and position-to-term rule. (5-6) Quadratic Sequences
More informationVirtual Vector Machine for Bayesian Online Classification
Virtual Vector Machine for Bayesian Online Classification Thomas P. Minka, Rongjing Xiang, Yuan (Alan) Qi Appeared in UAI 2009 Presented by Lingbo Li Introduction Online Learning Update model parameters
More informationCS 188: Artificial Intelligence Fall Machine Learning
CS 188: Artificial Intelligence Fall 2007 Lecture 23: Naïve Bayes 11/15/2007 Dan Klein UC Berkeley Machine Learning Up till now: how to reason or make decisions using a model Machine learning: how to select
More informationCONDITION CONTROLLED LOOPS. Introduction to Programming - Python
CONDITION CONTROLLED LOOPS Introduction to Programming - Python Generating Random Numbers Generating a random integer Sometimes you need your program to generate information that isn t available when you
More informationUsing Machine Learning to Optimize Storage Systems
Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation
More informationVoI for adversarial information structures
ARO ARO MURI MURI on on Value-centered Theory for for Adaptive Learning, Inference, Tracking, and and Exploitation VoI for adversarial information structures Kickf Emre Ertin, Nithin Sugavanam The Ohio
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 2: Probability: Discrete Random Variables Classification: Validation & Model Selection Many figures
More informationCS242: Probabilistic Graphical Models Lecture 3: Factor Graphs & Variable Elimination
CS242: Probabilistic Graphical Models Lecture 3: Factor Graphs & Variable Elimination Instructor: Erik Sudderth Brown University Computer Science September 11, 2014 Some figures and materials courtesy
More informationProbabilistic Learning Classification using Naïve Bayes
Probabilistic Learning Classification using Naïve Bayes Weather forecasts are usually provided in terms such as 70 percent chance of rain. These forecasts are known as probabilities of precipitation reports.
More informationWireless Network Security : Spring Arjun Athreya March 3, 2011 Survey: Trust Evaluation
Wireless Network Security 18-639: Spring 2011 Arjun Athreya March 3, 2011 Survey: Trust Evaluation A scenario LOBOS Management Co A CMU grad student new to Pittsburgh is looking for housing options in
More informationDetection and Mitigation of Cyber-Attacks using Game Theory
Detection and Mitigation of Cyber-Attacks using Game Theory João P. Hespanha Kyriakos G. Vamvoudakis Correlation Engine COAs Data Data Data Data Cyber Situation Awareness Framework Mission Cyber-Assets
More informationCS 188: Artificial Intelligence Fall Announcements
CS 188: Artificial Intelligence Fall 2006 Lecture 22: Naïve Bayes 11/14/2006 Dan Klein UC Berkeley Announcements Optional midterm On Tuesday 11/21 in class Review session 11/19, 7-9pm, in 306 Soda Projects
More informationAnnouncements. CS 188: Artificial Intelligence Fall Machine Learning. Classification. Classification. Bayes Nets for Classification
CS 88: Artificial Intelligence Fall 00 Lecture : Naïve Bayes //00 Announcements Optional midterm On Tuesday / in class Review session /9, 7-9pm, in 0 Soda Projects. due /. due /7 Dan Klein UC Berkeley
More informationDecentralized Stochastic Planning for Nonparametric Bayesian Models
Decentralized Stochastic Planning for Nonparametric Bayesian Models Silvia Ferrari Professor of Engineering and Computer Science Department of Mechanical Engineering and Materials Science Duke University
More informationChapter 5: Data compression. Chapter 5 outline
Chapter 5: Data compression Chapter 5 outline 2 balls weighing problem Examples of codes Kraft inequality Optimal codes + bounds Kraft inequality for uniquely decodable codes Huffman codes Shannon-Fano-Elias
More informationThe PageRank Computation in Google, Randomized Algorithms and Consensus of Multi-Agent Systems
The PageRank Computation in Google, Randomized Algorithms and Consensus of Multi-Agent Systems Roberto Tempo IEIIT-CNR Politecnico di Torino tempo@polito.it This talk The objective of this talk is to discuss
More informationMeshkat Farrokhzadi. in Partial Fulfillment of the Requirements for the Degree of. May 25, Certified b rvisor
Entropy, Information Rate and Mutual Information Measures for the Email Content of Information Workers by Meshkat Farrokhzadi Submitted to the Department of Electrical Engineering and Computer Science
More informationSAR change detection based on Generalized Gamma distribution. divergence and auto-threshold segmentation
SAR change detection based on Generalized Gamma distribution divergence and auto-threshold segmentation GAO Cong-shan 1 2, ZHANG Hong 1*, WANG Chao 1 1.Center for Earth Observation and Digital Earth, CAS,
More informationBayesian model selection and diagnostics
Bayesian model selection and diagnostics A typical Bayesian analysis compares a handful of models. Example 1: Consider the spline model for the motorcycle data, how many basis functions? Example 2: Consider
More informationLecture 1: An Introduction to Online Algorithms
Algoritmos e Incerteza (PUC-Rio INF979, 017.1) Lecture 1: An Introduction to Online Algorithms Mar 1, 017 Lecturer: Marco Molinaro Scribe: Joao Pedro T. Brandao Online algorithms differ from traditional
More informationLexical Semantics. Regina Barzilay MIT. October, 5766
Lexical Semantics Regina Barzilay MIT October, 5766 Last Time: Vector-Based Similarity Measures man woman grape orange apple n Euclidian: x, y = x y = i=1 ( x i y i ) 2 n x y x i y i i=1 Cosine: cos( x,
More informationAI Programming CS S-15 Probability Theory
AI Programming CS662-2013S-15 Probability Theory David Galles Department of Computer Science University of San Francisco 15-0: Uncertainty In many interesting agent environments, uncertainty plays a central
More informationContents. 3 Vector Quantization The VQ Advantage Formulation Optimality Conditions... 48
Contents Part I Prelude 1 Introduction... 3 1.1 Audio Coding... 4 1.2 Basic Idea... 6 1.3 Perceptual Irrelevance... 8 1.4 Statistical Redundancy... 9 1.5 Data Modeling... 9 1.6 Resolution Challenge...
More informationCS154, Lecture 18: 1
CS154, Lecture 18: 1 CS 154 Final Exam Wednesday December 13, 12:15-3:15 pm Skilling Auditorium You re allowed one double-sided sheet of notes Exam is comprehensive (but will emphasize post-midterm topics)
More informationround decimals to the nearest decimal place and order negative numbers in context
6 Numbers and the number system understand and use proportionality use the equivalence of fractions, decimals and percentages to compare proportions use understanding of place value to multiply and divide
More informationLecture 17. Lower bound for variable-length source codes with error. Coding a sequence of symbols: Rates and scheme (Arithmetic code)
Lecture 17 Agenda for the lecture Lower bound for variable-length source codes with error Coding a sequence of symbols: Rates and scheme (Arithmetic code) Introduction to universal codes 17.1 variable-length
More informationCrypto-systems all around us ATM machines Remote logins using SSH Web browsers (https invokes Secure Socket Layer (SSL))
Introduction (Mihir Bellare Text/Notes: http://cseweb.ucsd.edu/users/mihir/cse207/) Cryptography provides: Data Privacy Data Integrity and Authenticity Crypto-systems all around us ATM machines Remote
More informationFOUNDATION FOR DATA RESOURCE QUALITY
FOUNDATION FOR DATA RESOURCE QUALITY CENTRAL OHIO DAMA APRIL 14, 2005 MICHAEL H. BRACKETT CONSULTING DATA ARCHITECT DATA RESOURCE DESIGN & REMODELING DRDR P.O. Box 151 Lilliwaup, Washington 98555 Voice:
More informationName. Final Exam, Economics 210A, December 2012 There are 8 questions. Answer as many as you can... Good luck!
Name Final Exam, Economics 210A, December 2012 There are 8 questions. Answer as many as you can... Good luck! 1) Let S and T be convex sets in Euclidean n space. Let S + T be the set {x x = s + t for some
More informationDeep Learning Cook Book
Deep Learning Cook Book Robert Haschke (CITEC) Overview Input Representation Output Layer + Cost Function Hidden Layer Units Initialization Regularization Input representation Choose an input representation
More informationBayesian Classification Using Probabilistic Graphical Models
San Jose State University SJSU ScholarWorks Master's Projects Master's Theses and Graduate Research Spring 2014 Bayesian Classification Using Probabilistic Graphical Models Mehal Patel San Jose State University
More informationAdversary Arguments. A method for obtaining lower bounds
Adversary Arguments A method for obtaining lower bounds What is an Adversary? A Second Algorithm Which Intercepts Access to Data Structures Constructs the input data only as needed Attempts to make original
More informationThe PageRank Computation in Google, Randomized Algorithms and Consensus of Multi-Agent Systems
The PageRank Computation in Google, Randomized Algorithms and Consensus of Multi-Agent Systems Roberto Tempo IEIIT-CNR Politecnico di Torino tempo@polito.it This talk The objective of this talk is to discuss
More informationLecture 15. Error-free variable length schemes: Shannon-Fano code
Lecture 15 Agenda for the lecture Bounds for L(X) Error-free variable length schemes: Shannon-Fano code 15.1 Optimal length nonsingular code While we do not know L(X), it is easy to specify a nonsingular
More informationThresholds Determination for Probabilistic Rough Sets with Genetic Algorithms
Thresholds Determination for Probabilistic Rough Sets with Genetic Algorithms Babar Majeed, Nouman Azam, JingTao Yao Department of Computer Science University of Regina {majeed2b,azam200n,jtyao}@cs.uregina.ca
More informationAlternatives to Direct Supervision
CreativeAI: Deep Learning for Graphics Alternatives to Direct Supervision Niloy Mitra Iasonas Kokkinos Paul Guerrero Nils Thuerey Tobias Ritschel UCL UCL UCL TUM UCL Timetable Theory and Basics State of
More informationGradient Descent. Wed Sept 20th, James McInenrey Adapted from slides by Francisco J. R. Ruiz
Gradient Descent Wed Sept 20th, 2017 James McInenrey Adapted from slides by Francisco J. R. Ruiz Housekeeping A few clarifications of and adjustments to the course schedule: No more breaks at the midpoint
More informationExtreme Value Theory in (Hourly) Precipitation
Extreme Value Theory in (Hourly) Precipitation Uli Schneider Geophysical Statistics Project, NCAR GSP Miniseries at CSU November 17, 2003 Outline Project overview Extreme value theory 101 Applying extreme
More informationCAP 5993/CAP 4993 Game Theory. Instructor: Sam Ganzfried
CAP 5993/CAP 4993 Game Theory Instructor: Sam Ganzfried sganzfri@cis.fiu.edu 1 Announcements HW 1 due today HW 2 out this week (2/2), due 2/14 2 Definition: A two-player game is a zero-sum game if for
More informationComputational Cognitive Science
Computational Cognitive Science Lecture 5: Maximum Likelihood Estimation; Parameter Uncertainty Chris Lucas (Slides adapted from Frank Keller s) School of Informatics University of Edinburgh clucas2@inf.ed.ac.uk
More informationLAB 5: SELECTION STATEMENTS
Statement Purpose: The purpose of this lab is to familiarize students with Conditional statements and explain how to control the sequence of statement execution, depending on the value of an expression
More informationINLA: Integrated Nested Laplace Approximations
INLA: Integrated Nested Laplace Approximations John Paige Statistics Department University of Washington October 10, 2017 1 The problem Markov Chain Monte Carlo (MCMC) takes too long in many settings.
More information1.1 calculator viewing window find roots in your calculator 1.2 functions find domain and range (from a graph) may need to review interval notation
1.1 calculator viewing window find roots in your calculator 1.2 functions find domain and range (from a graph) may need to review interval notation functions vertical line test function notation evaluate
More informationCommunication Amid Uncertainty
Communication Amid Uncertainty Madhu Sudan Microsoft Research Based on Juba, S. (STOC 2008, ITCS 2011) Goldreich, Juba, S. (JACM 2011) Juba, Kalai, Khanna, S. (ITCS 2011) Haramaty, S. (ITCS 2014) Canonne,
More informationDigital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay
Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 29 Source Coding (Part-4) We have already had 3 classes on source coding
More informationExam Topics. Search in Discrete State Spaces. What is intelligence? Adversarial Search. Which Algorithm? 6/1/2012
Exam Topics Artificial Intelligence Recap & Expectation Maximization CSE 473 Dan Weld BFS, DFS, UCS, A* (tree and graph) Completeness and Optimality Heuristics: admissibility and consistency CSPs Constraint
More informationDigital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay
Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 26 Source Coding (Part 1) Hello everyone, we will start a new module today
More informationA Well-Behaved Algorithm for Simulating Dependence Structures of Bayesian Networks
A Well-Behaved Algorithm for Simulating Dependence Structures of Bayesian Networks Yang Xiang and Tristan Miller Department of Computer Science University of Regina Regina, Saskatchewan, Canada S4S 0A2
More informationClustering Using Elements of Information Theory
Clustering Using Elements of Information Theory Daniel de Araújo 1,2, Adrião Dória Neto 2, Jorge Melo 2, and Allan Martins 2 1 Federal Rural University of Semi-Árido, Campus Angicos, Angicos/RN, Brasil
More informationCommunication Amid Uncertainty
Communication Amid Uncertainty Madhu Sudan Microsoft Research Based on Juba, S. (STOC 2008, ITCS 2011) Goldreich, Juba, S. (JACM 2011) Juba, Kalai, Khanna, S. (ITCS 2011) Haramaty, S. (ITCS 2014) Canonne,
More informationWhat is a Computer Algorithm?
What is a Computer Algorithm? A computer algorithm is a detailed step-by-step method for solving a problem by using a computer. Problem-Solving (Science and Engineering) Analysis How does it work? Breaking
More informationComputer security experts often tell us not to choose an easy to guess word as a password. For example, the user name or date of birth.
Graphical Passwords: Computer security experts often tell us not to choose an easy to guess word as a password. For example, the user name or date of birth. The Graphical Passwords team have been working
More informationEE67I Multimedia Communication Systems Lecture 4
EE67I Multimedia Communication Systems Lecture 4 Lossless Compression Basics of Information Theory Compression is either lossless, in which no information is lost, or lossy in which information is lost.
More informationRandomized algorithms have several advantages over deterministic ones. We discuss them here:
CS787: Advanced Algorithms Lecture 6: Randomized Algorithms In this lecture we introduce randomized algorithms. We will begin by motivating the use of randomized algorithms through a few examples. Then
More informationDavid Rappaport School of Computing Queen s University CANADA. Copyright, 1996 Dale Carnegie & Associates, Inc.
David Rappaport School of Computing Queen s University CANADA Copyright, 1996 Dale Carnegie & Associates, Inc. Data Compression There are two broad categories of data compression: Lossless Compression
More informationCopyright 2000, Kevin Wayne 1
Guessing Game: NP-Complete? 1. LONGEST-PATH: Given a graph G = (V, E), does there exists a simple path of length at least k edges? YES. SHORTEST-PATH: Given a graph G = (V, E), does there exists a simple
More informationMontana City School GRADE 5
Montana City School GRADE 5 Montana Standard 1: Students engage in the mathematical processes of problem solving and reasoning, estimation, communication, connections and applications, and using appropriate
More informationArtificial Intelligence
Artificial Intelligence Other models of interactive domains Marc Toussaint University of Stuttgart Winter 2018/19 Basic Taxonomy of domain models Other models of interactive domains Basic Taxonomy of domain
More informationWelcome to the course Algorithm Design
Welcome to the course Algorithm Design Summer Term 2011 Friedhelm Meyer auf der Heide Lecture 12, 8.7.2011 Friedhelm Meyer auf der Heide 1 Randomised Algorithms Friedhelm Meyer auf der Heide 2 Topics -
More informationList of NEW Maths content
List of NEW Maths content Our brand new Maths content for the new Maths GCSE (9-1) consists of 212 chapters broken up into 37 titles and 4 topic areas (Algebra, Geometry & Measures, Number and Statistics).
More informationPerformance Level Descriptors. Mathematics
Performance Level Descriptors Grade 3 Well Students rarely, Understand that our number system is based on combinations of 1s, 10s, and 100s (place value, compare, order, decompose, and combine using addition)
More informationFinal Exam. Introduction to Artificial Intelligence. CS 188 Spring 2010 INSTRUCTIONS. You have 3 hours.
CS 188 Spring 2010 Introduction to Artificial Intelligence Final Exam INSTRUCTIONS You have 3 hours. The exam is closed book, closed notes except a two-page crib sheet. Please use non-programmable calculators
More informationIntroduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao
Introduction to GAN Generative Adversarial Networks Junheng(Jeff) Hao Adversarial Training is the coolest thing since sliced bread. -- Yann LeCun Roadmap 1. Generative Modeling 2. GAN 101: What is GAN?
More informationSupervised vs unsupervised clustering
Classification Supervised vs unsupervised clustering Cluster analysis: Classes are not known a- priori. Classification: Classes are defined a-priori Sometimes called supervised clustering Extract useful
More informationExpectation-Maximization Methods in Population Analysis. Robert J. Bauer, Ph.D. ICON plc.
Expectation-Maximization Methods in Population Analysis Robert J. Bauer, Ph.D. ICON plc. 1 Objective The objective of this tutorial is to briefly describe the statistical basis of Expectation-Maximization
More informationComputer Vision. Exercise Session 10 Image Categorization
Computer Vision Exercise Session 10 Image Categorization Object Categorization Task Description Given a small number of training images of a category, recognize a-priori unknown instances of that category
More informationCDG2A/CDZ4A/CDC4A/ MBT4A ELEMENTS OF OPERATIONS RESEARCH. Unit : I - V
CDG2A/CDZ4A/CDC4A/ MBT4A ELEMENTS OF OPERATIONS RESEARCH Unit : I - V UNIT I Introduction Operations Research Meaning and definition. Origin and History Characteristics and Scope Techniques in Operations
More informationChapter 1 Polynomials and Modeling
Chapter 1 Polynomials and Modeling 1.1 Linear Functions Recall that a line is a function of the form y = mx+ b, where m is the slope of the line (how steep the line is) and b gives the y-intercept (where
More informationCSE 421 Greedy: Huffman Codes
CSE 421 Greedy: Huffman Codes Yin Tat Lee 1 Compression Example 100k file, 6 letter alphabet: File Size: ASCII, 8 bits/char: 800kbits 2 3 > 6; 3 bits/char: 300kbits a 45% b 13% c 12% d 16% e 9% f 5% Why?
More informationMinnesota Academic Standards for Mathematics 2007
An Alignment of Minnesota for Mathematics 2007 to the Pearson Integrated High School Mathematics 2014 to Pearson Integrated High School Mathematics Common Core Table of Contents Chapter 1... 1 Chapter
More informationMonte Carlo Methods and Statistical Computing: My Personal E
Monte Carlo Methods and Statistical Computing: My Personal Experience Department of Mathematics & Statistics Indian Institute of Technology Kanpur November 29, 2014 Outline Preface 1 Preface 2 3 4 5 6
More informationMath Interim Mini-Tests. 3rd Grade Mini-Tests
3rd Grade Mini-Tests Mini-Test Name Availability Area of Plane Figures-01 Gr 3_Model, Reason, & Solve Problems-04 Multiplicative Properties & Factors-01 Patterns & Using the Four Operations-01 Real-World
More informationUnsupervised Learning
Deep Learning for Graphics Unsupervised Learning Niloy Mitra Iasonas Kokkinos Paul Guerrero Vladimir Kim Kostas Rematas Tobias Ritschel UCL UCL/Facebook UCL Adobe Research U Washington UCL Timetable Niloy
More informationAlgorithms for Markov Random Fields in Computer Vision
Algorithms for Markov Random Fields in Computer Vision Dan Huttenlocher November, 2003 (Joint work with Pedro Felzenszwalb) Random Field Broadly applicable stochastic model Collection of n sites S Hidden
More information