Lecture Outlines. CSCI- 360 Section R Spring 2016 Wei- Min Shen

Similar documents
CSCI 360 Introduc/on to Ar/ficial Intelligence Week 2: Problem Solving and Op/miza/on. Instructor: Wei-Min Shen

Exam Topics. Search in Discrete State Spaces. What is intelligence? Adversarial Search. Which Algorithm? 6/1/2012

Probabilistic Belief. Adversarial Search. Heuristic Search. Planning. Probabilistic Reasoning. CSPs. Learning CS121

10-701/15-781, Fall 2006, Final

Markov Decision Processes and Reinforcement Learning

CSCI-630 Foundations of Intelligent Systems Fall 2015, Prof. Zanibbi

CS 416, Artificial Intelligence Midterm Examination Fall 2004

Unsupervised Learning: Clustering

Artificial Intelligence. Chapters Reviews. Readings: Chapters 3-8 of Russell & Norvig.

Data Structures and Algorithms

CS 188: Artificial Intelligence Spring Recap Search I

Recap Search I. CS 188: Artificial Intelligence Spring Recap Search II. Example State Space Graph. Search Problems.

7 Temporal Models (20 points)

Potential Midterm Exam Questions

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

Using Machine Learning to Optimize Storage Systems

CSEP 573: Artificial Intelligence

CS 540: Introduction to Artificial Intelligence

Graphical Models. David M. Blei Columbia University. September 17, 2014

Introduction to Graphical Models

CS-171, Intro to A.I. Mid-term Exam Fall Quarter, 2013

Classification Lecture Notes cse352. Neural Networks. Professor Anita Wasilewska

ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

Final Exam. Introduction to Artificial Intelligence. CS 188 Spring 2010 INSTRUCTIONS. You have 3 hours.

Markov Decision Processes (MDPs) (cont.)

Definition: A context-free grammar (CFG) is a 4- tuple. variables = nonterminals, terminals, rules = productions,,

Practice EXAM: SPRING 2012 CS 6375 INSTRUCTOR: VIBHAV GOGATE

Name: CSCI E-220: Artificial Intelligence Midterm Exam, Fall Term 2001

ECG782: Multidimensional Digital Signal Processing

Finding optimal configurations Adversarial search

Planning and Control: Markov Decision Processes

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas

Uninformed Search Methods. Informed Search Methods. Midterm Exam 3/13/18. Thursday, March 15, 7:30 9:30 p.m. room 125 Ag Hall

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov

18 October, 2013 MVA ENS Cachan. Lecture 6: Introduction to graphical models Iasonas Kokkinos

Question 1: 25% of students lost more than 2 points Question 2: 50% of students lost 2-4 points; 50% got it entirely correct Question 3: 95% of

Search and Optimization

Artificial Intelligence. Programming Styles

CS 188: Artificial Intelligence. Recap Search I

Propositional logic (Ch. 7)

K-Means and Gaussian Mixture Models

CS-171, Intro to A.I. Mid-term Exam Summer Quarter, 2016

Reinforcement Learning and Optimal Control. ASU, CSE 691, Winter 2019

Dynamic Bayesian network (DBN)

CS 540: Introduction to Artificial Intelligence

More Learning. Ensembles Bayes Rule Neural Nets K-means Clustering EM Clustering WEKA

Knowledge Representation. CS 486/686: Introduction to Artificial Intelligence

Modeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA

Artificial Intelligence

Unsupervised Learning. CS 3793/5233 Artificial Intelligence Unsupervised Learning 1

CSE 573: Artificial Intelligence Autumn 2010

Machine Learning Classifiers and Boosting

CS-171, Intro to A.I. Mid-term Exam Winter Quarter, 2014

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning

CSC242: Intro to AI. Lecture 8 LOGIC

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves

Planning: Wrap up CSP Planning Logic: Intro

Midterm Examination CS 540-2: Introduction to Artificial Intelligence

Artificial Intelligence (part 4a) Problem Solving Using Search: Structures and Strategies for State Space Search

Reinforcement Learning: A brief introduction. Mihaela van der Schaar

To earn the extra credit, one of the following has to hold true. Please circle and sign.

Notes for Chapter 12 Logic Programming. The AI War Basic Concepts of Logic Programming Prolog Review questions

The Basics of Graphical Models

7. Boosting and Bagging Bagging

Tillämpad Artificiell Intelligens Applied Artificial Intelligence Tentamen , , MA:9

Machine Learning

Name: UW CSE 473 Midterm, Fall 2014

Logic: TD as search, Datalog (variables)

Two-player Games ZUI 2016/2017

Review Agents ( ) Review State Space Search

CS-171, Intro to A.I. Mid-term Exam Fall Quarter, 2014

CSC242: Intro to AI. Lecture 10. Tuesday, February 26, 13

Partially Observable Markov Decision Processes. Silvia Cruciani João Carvalho

(Due to rounding, values below may be only approximate estimates.) We will supply these numbers as they become available.

Algorithm Design Techniques. Hwansoo Han

Applied Statistics for Neuroscientists Part IIa: Machine Learning

4 Search Problem formulation (23 points)

Network Traffic Measurements and Analysis

Artificial Intelligence

Probabilistic Graphical Models

Object Recognition. Lecture 11, April 21 st, Lexing Xie. EE4830 Digital Image Processing

You should provide the best solutions you can think of.

CS 4700: Artificial Intelligence

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS

Grad AI Fall, 2007 Homework 1

Learning. Learning agents Inductive learning. Neural Networks. Different Learning Scenarios Evaluation

Supervised Learning with Neural Networks. We now look at how an agent might learn to solve a general problem by seeing examples.

Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound. Lecture 12: Deep Reinforcement Learning

CS570: Introduction to Data Mining

Search Engines. Information Retrieval in Practice

Chapter 2 Classical algorithms in Search and Relaxation

Announcements. Homework 4. Project 3. Due tonight at 11:59pm. Due 3/8 at 4:00pm

Lecture 21 : A Hybrid: Deep Learning and Graphical Models

COMP90051 Statistical Machine Learning

Introduction to Machine Learning CMU-10701

CPSC 436D Video Game Programming

CS-171, Intro to A.I. Mid-term Exam Winter Quarter, 2016

CS 229 Midterm Review

Chapter 3: Search. c D. Poole, A. Mackworth 2010, W. Menzel 2015 Artificial Intelligence, Chapter 3, Page 1

COS Lecture 13 Autonomous Robot Navigation

Transcription:

Lecture Outlines CSCI- 360 Section- 30304R Spring 2016 Wei- Min Shen

Wk1 Ch1: Introduction * Welcome * Class structures/rules (syllabus, HM, projects, tests) * What is AI? * What are the related fields of AI? * How was AI started? (history) * What is the state of the art of AI?

Wk1 Ch2: Intelligent Agents * Show SuperBot movies and simulations * Project 1: Description and assignment * Agent/Robot * What is inside a robot/agent? (see, act, think, learn) * What does it seek? (goal, utility, solve problems, fame/ fortune ) * Environment/world * What can be seen? * What can be changed? * Who else are there? (Obstacle and other agents) * HW: Define an environment/world and a robot by yourself

Wk2 Ch3: Solving Problems * How to represent a problem? * states, actions, initials, goals, (e.g., tic- tac- toe) * How to solve a problem? * Search: from here to there, initials to goals * Depth- first, breadth- first * How good is your solution? (fig 6.1, ALFE) * How good is your state? How costly is an action? * Best- first, Dynamic Programming, A*, etc. * Can you guarantee anything? (optimal vs heuristic) * How much do you want to pay for your solution? * How deep/wide can you go? * Predetermined vs dynamic (e.g., iterative deepening) * One way or many ways (bi- directional)? * How big is a problem? Can you put the whole world in your head? * Tower of Hanoi, chess, robot- and- world, * HM: state space for TOH, assign values for state and actions.

Wk2 Ch4: Optimization * How to find the extremes of f(x)? * Why is it so important? * Why is it so hard? (no one has offered a general solution!) * Which way to go? * How much can you see? (local vs global) * How many points can you remember? * How big is your step? (skip THE point?) * How well can you guess? * How much do you know about the function? * How do you know you are done? * Will the function change by itself? * HW: your own answers for the above questions

Wk3 Ch5: Game Playing * Key idea: optimization for youself and your opponent * When you can see the whole game * Example, Tic- tac- toe (fig 5.1) * Optimal: Can I always win, or never loss? * Key: look ahead, chose the best * Minmax algorithm, Alpha- beta pruning * Fig 5.2, key idea: what can I gain if the opponent gives me the least * When cannot see the whole game (chess) * how far to look? how good is a state? Can you predict your opponent? * When there is a random player (dice) * Is random playing any good? * What are the best AI game players? * HW: Write a minmax algorithm for tic- tac- toe

Wk3 Ch6: Constraint Satisfaction * Example: Map Coloring, can you think of another? * Formalism: Variables, Value Sets, Constraints, * Independent subset? Existing cycles? * The basic algorithms * Global Backtracking * Order variables: the most constrained variables first * Satisfy one at a time, if stuck, backtracking * Local search: heuristic, find the min- conflicts solution locally * Centralized vs. distributed (DCOP) * HW: Magic square, define it and solve it

Wk4 Ch7-9: Logics * The real world and its representation by logics * The world you created in simulation (e.g. wumpus) * syntax(on paper, α,β),,,, * semantics(in world/model M), satisfy * Propositional logic: truth table for * Theorem proving, inference rule, resolution, horn clause, forward/backward chaining, * Propositional agent: current state, goal state, local search, global search * HW: PL rules for Wumpus or Tic- Tac- Toe α = β M(α ) M(β)

Wk4 Ch8: First- order logic * Syntax and semantics * Add variables (predicates) and quantifiers * Knowledge engineering * Capture and represent human knowledge, e.g., using FOL * HW: FOL rules for Wumpus or Tic- Tac- Toe

Wk4 Ch9: Inference in FOL * Modus Ponens with variables * Unification: finding the values for variables * 1. Forward chaining * E.g. production systems, ACT, SOAR * 2. Backward chaining * E.g., Prolog * 3. Resolution * Convert to CNF, eliminating P, - P.

Wk5 Ch10: Actions and Planning * Project 2: Robot searches a target room on a map * Representing Actions * Pre and post condition (single or set of states) * Searching for optimal solutions * Graph- Based, Forward, Backward. * Dynamic Programming, A*, Q- search (read ALFE 6.1) * Searching for satisfacing solutions * means- end- analysis * Heuristic planning * Other planning methods * Logic deduction: Boolean satisfactory, situation- calculus * Constraint- satisfaction * Partially ordered plans

Wk5 Ch11: Scheduling * Why hard? Time, duration, resources, deadlines, orders, unknown effects, etc. * Deterministic approaches * The longest first (critical path, take at least this long) * Hierarchical Planning * High- level actions first * Find primitive solutions * Find abstract solutions * Non- deterministic approaches * Sensor- based, contingency, incremental, * Multi- agent planning (collaboration & coordination) * HW: modify Wumpus as a scheduling problem

Wk6 Ch12: Knowledge Representation * What is KR? Representation + Processes * Why important? Simon s example cards/ttt * Format: Attribute- based and Relational- based * Ontology * Objects: (attributes, part- of), both physical and mental * Events (process, time), * Relations, functions, operators: mostly mental * Reasoning * Semantic net, logic, default (monotonic or not), truth maintenance * Example * Work out an ontology for the Wumpus based on the example of Internet shopping in the book

Wk6: Logics and Probabilities * Not all knowledge are certain * Earlier stories of expert systems (medical, legal) * Two big challenges * Common sense (vague): Water flows down * Uncertainty * Probability =?= Logic of Science

Wk6-7 Ch13: Probability * Where is probability? (in the world or in your mind) * Notation: variable X, value x i, P(X=x i ), P(X) denotes for all values of X, P(X,Y), P(X,y) * Remember two axioms * Sum axiom: P(A B)+P(~A B)=1 * Product axiom: P(AB C)=P(A C)P(B AC)=P(B C)P(A BC) * They can do much more than logics * Deductive reasoning: * If A- >B and A, then B * If A- >B and ~B, then ~A * If A- >B and B, then A become more plausible * Inductive reasoning: * If A- >B and ~A, then B become less plausible * If A- > B becomes more plausible and B, then A become more plausible * HM: work out the math why the above is true (ALFE, p102)

Wk7 Ch14: Probability Reasoning * Propositional logic + probability * Full joint distribution (fig 13.3: truth table + probability) * Inferences * Using full joint distributions (1 st part of product rule) * Marginalization or summing out * Using Bayesian rule (2 nd part of product rule) * Example for Wumpus world * Bayesian network: topology + CPTs (fig 14.2) * Compare to truth- table format (2 5 rows) * Inferences, exact vs approximate * Relational and FOL + probability * Other uncertain reasoning theories * Dempster- Shafer, Fuzzy sets and logic,

Wk8: HMM, DBN, Automata * Readings: AIMA 15 * Examples: Speech recognition, see =?= truth * Given: a map of colored rooms, and an experience (act/see): e 1:t * Compute: which room the robot was/is/will- be in: X * States, observations, transitions, actions, sensors, * Definitions (ALFE 5.10) * We can infer: * P(X t e 1:t ) which room I am in now (state estimation) * P(X t+k e 1:t ) which room I will be in at time t+k (prediction) * P(X k e 1:t ) which room I was in at time k (smoothing) * P(x 1:t e 1:t ) what rooms I have been in the past (explain) * P(M t e 1:t ) can I improve my map (learning) * HW: Define an automata for your rooms in the simulation

Wk9: Utility, Actions, MDP, POMDP * Readings: AIMA 16, 17 * States have utility U(s) * Maximum Expected Utility * EU(a e) = Sum[P(s a,e)u(s)] * Markov Decision Process (MDP): state, action, transition P(s s,a), reward: sè r, or (s,a)è r * Solutions are represented as policy: sè a. * Total discounted expected reward (Bellman): * Equation 17.5 or in ALFE page 167. * Partially Observed MDP (POMDP) * States cannot be completely observed * Need a sensor model P(z s) * Objective: finding the optimal policy based on utilities * Review for Midterm exam (if you have any questions) * HW: Define your own world and robot using POMDP

* Good luck on your test! Wk10: Midterm

Wk11: Supervised Learning 1: Attribute- Based * Types Readings: AIMA Ch18, ALFE 4 * Project 3: a robot learns a model of a colored room to find a target quickly * Types of Machine Learning * Unsupervised vs. Supervised (examples, reinforcement, and environment) * Incremental vs. patch; online vs. offline * Key Ideas: Hypotheses and examples * Hypotheses space (ALFE fig4.3): Attribute- based vs relation- based * Learning = Search in hypothesis space for one to cover examples (both seen and unseen) * False negative (cover too little) à Generalization * False positive (cover too much) à Specialization * Correctness criteria: Overfitting, Identical, PAC, and others * Algorithms for Attribute- Based Learning * Version Space * Decision Trees (HW: ) * Neural Network (fig 18.19 and ALEF 4.11) (HW:) * Linear Regression: Learn h=ax+b from examples (x,y) by minimizing Σ(y- h) 2 [eq 18.2-3] * Support vector machines (HW: 18.16, 18.25) * Ensemble Learning and Boosting * Learn a new hypothesis based on the results of the last hypothesis (weight the examples)

Wk12: Supervised Learning 2 Relation- Based * Readings: AIMA 19, ALFE 4.7-4.10 * A general challenge: a good hypothesis space * Too open: too slow to converge * Too restrict: rule out the good hypotheses * Good priori knowledge would help * Relation- based learning * Deductive (e.g. EBL): use priori knowledge to explain examples à new hypotheses (fig 19.7) * Inductive algorithms (ALFE 4.7-4.10) * FOIL * CDL (Complementary Discrimination Learning) * HW: learn the relational rules of X from examples

Wk13: Unsupervised Learning Probability- Based * Readings: AIMA 20, ALFE 5.10 * Clustering * Naïve Bayes Models * By Bayesian Approach (Cheeseman s CLASSIFY) * K- mean algorithm (see Andrew Ng s lecture) * The EM algorithm * Clustering with mixtures of Gaussians (fig 20.11) * Bayesian network with hidden variables (fig 20.13) * Hidden Markov models (fig 20.14) * PO- MDP (ALFE 5.10) * HW: Learn a Bayesian model from examples

Week13: Reinforcement Learning * Reinforcement Learning * Delayed supervision/reward * When we only know the rewards from the goals

Wk14: Surprise- Based Learning * Lecture by Dr. Nadeesha Ranasinghe

Wk15: Integrated Intelligence * Actions and Perceptions * Communication * Natural language overview * Speech understanding * Collaboration * Self- Organization and self- reconfiguration * Rubenstein 2010.

Mk16: Towards the Future * Relational + Probabilistic Models * Existing research: FOL + Probability * Structural Learning * Existing research: SBL, * Scale up to the real world * Integrate with all aspects * Reconfigure both mind and body

* Good Luck! Mk17: Final Exam

* See following slides Extra Lectures

Graphics a 0 Room- x a 1 Room- y a 2 Room- z a 3 Wall- color Wall- color Wall- color

Support Vector Machines * Support vector machines (HW: 18.16, 18.25) * The key idea of max- margin separator (fig 18.30) * Support vectors: the points closest to the separator, all other data points have zero weights * When the data are not linearly separable (fig 18.31), rewrite separator in a higher dimension * It is non- parametric, but (eq. 18.13) can use fewer examples than the training data