Uninformed Search Methods. Informed Search Methods. Midterm Exam 3/13/18. Thursday, March 15, 7:30 9:30 p.m. room 125 Ag Hall

Similar documents
Midterm Examination CS 540-2: Introduction to Artificial Intelligence

Midterm Examination CS540-2: Introduction to Artificial Intelligence

CS 540: Introduction to Artificial Intelligence

University of Waterloo Department of Electrical and Computer Engineering ECE 457A: Cooperative and Adaptive Algorithms Midterm Examination

CSCI-630 Foundations of Intelligent Systems Fall 2015, Prof. Zanibbi

CS 188: Artificial Intelligence. Recap Search I

Lecture 27: Review. Reading: All chapters in ISLR. STATS 202: Data mining and analysis. December 6, 2017

DIT411/TIN175, Artificial Intelligence. Peter Ljunglöf. 23 January, 2018

Last time: Problem-Solving

Advanced A* Improvements

Lecture 4: Search 3. Victor R. Lesser. CMPSCI 683 Fall 2010

Downloaded from ioenotes.edu.np

Artificial Intelligence

Heuristic (Informed) Search

3.6.2 Generating admissible heuristics from relaxed problems

Artificial Intelligence

CS 540-1: Introduction to Artificial Intelligence

Potential Midterm Exam Questions

Local Search and Optimization Chapter 4. Mausam (Based on slides of Padhraic Smyth, Stuart Russell, Rao Kambhampati, Raj Rao, Dan Weld )

Local Search and Optimization Chapter 4. Mausam (Based on slides of Padhraic Smyth, Stuart Russell, Rao Kambhampati, Raj Rao, Dan Weld )

Gradient Descent. 1) S! initial state 2) Repeat: Similar to: - hill climbing with h - gradient descent over continuous space

Classification with Decision Tree Induction

Local Search and Optimization Chapter 4. Mausam (Based on slides of Padhraic Smyth, Stuart Russell, Rao Kambhampati, Raj Rao, Dan Weld )

(Due to rounding, values below may be only approximate estimates.) We will supply these numbers as they become available.

Chapters 3-5 Problem Solving using Search

N-Queens problem. Administrative. Local Search

CS-171, Intro to A.I. Mid-term Exam Winter Quarter, 2016

Artificial Intelligence

MIT 801. Machine Learning I. [Presented by Anna Bosman] 16 February 2018

Classification Using Genetic Programming. Patrick Kellogg General Assembly Data Science Course (8/23/15-11/12/15)

Artificial Intelligence. Chapters Reviews. Readings: Chapters 3-8 of Russell & Norvig.

mywbut.com Informed Search Strategies-II

Artificial Intelligence. Programming Styles

Searching. Assume goal- or utilitybased. Next task to achieve is to determine the best path to the goal

Artificial Intelligence

Using Machine Learning to Optimize Storage Systems

CS 380: Artificial Intelligence Lecture #4

Monotonicity. Admissible Search: That finds the shortest path to the Goal. Monotonicity: local admissibility is called MONOTONICITY

Monte Carlo Tree Search

Lecture 25: Review I

Local Search (Greedy Descent): Maintain an assignment of a value to each variable. Repeat:

Today s s lecture. Lecture 3: Search - 2. Problem Solving by Search. Agent vs. Conventional AI View. Victor R. Lesser. CMPSCI 683 Fall 2004

Review: Final Exam CPSC Artificial Intelligence Michael M. Richter

SRI VIDYA COLLEGE OF ENGINEERING & TECHNOLOGY REPRESENTATION OF KNOWLEDGE PART A

ARTIFICIAL INTELLIGENCE. Informed search

Informed Search and Exploration

CSC384 Midterm Sample Questions

CS-171, Intro to A.I. Mid-term Exam Fall Quarter, 2017

Administrative. Local Search!

Unsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi

Exam Topics. Search in Discrete State Spaces. What is intelligence? Adversarial Search. Which Algorithm? 6/1/2012

Review Adversarial (Game) Search ( ) Review Constraint Satisfaction ( ) Please review your quizzes and old CS-271 tests

CS 416, Artificial Intelligence Midterm Examination Fall 2004

Outline. Best-first search

Introduction to Machine Learning. Xiaojin Zhu

Heuristic Search. Heuristic Search. Heuristic Search. CSE 3401: Intro to AI & LP Informed Search

Program of Study. Artificial Intelligence 1. Shane Torbert TJHSST

Outline. Informed Search. Recall: Uninformed Search. An Idea. Heuristics Informed search techniques More on heuristics Iterative improvement

4 INFORMED SEARCH AND EXPLORATION. 4.1 Heuristic Search Strategies

Artificial Intelligence

CS 4700: Artificial Intelligence

ARTIFICIAL INTELLIGENCE LECTURE 3. Ph. D. Lect. Horia Popa Andreescu rd year, semester 5

Final Exam. Introduction to Artificial Intelligence. CS 188 Spring 2010 INSTRUCTIONS. You have 3 hours.

Outline. Best-first search

Spring 2007 Midterm Exam

Cluster Analysis: Agglomerate Hierarchical Clustering

Introduction to Spring 2007 Artificial Intelligence Midterm Solutions

Artificial Intelligence p.1/49. n-queens. Artificial Intelligence p.2/49. Initial state: the empty board or a board with n random

Name: UW CSE 473 Midterm, Fall 2014

Clustering. Robert M. Haralick. Computer Science, Graduate Center City University of New York

Unsupervised Learning

CS 4700: Artificial Intelligence

CS 188: Artificial Intelligence Spring Recap Search I

Data Mining Lecture 8: Decision Trees

Unsupervised Learning and Clustering

1 Tree Search (12 points)

Hill Climbing. Assume a heuristic value for each assignment of values to all variables. Maintain an assignment of a value to each variable.

Artificial Intelligence. Game trees. Two-player zero-sum game. Goals for the lecture. Blai Bonet

Introduction to Computer Science and Programming for Astronomers

Class Overview. Introduction to Artificial Intelligence COMP 3501 / COMP Lecture 2. Problem Solving Agents. Problem Solving Agents: Assumptions

Informed search algorithms. Chapter 4

Advanced Search Genetic algorithm

Solving Problems: Intelligent Search

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

CPSC 436D Video Game Programming

Class Overview. Introduction to Artificial Intelligence COMP 3501 / COMP Lecture 2: Search. Problem Solving Agents

Artificial Intelligence

Unsupervised Learning and Clustering

Recap Search I. CS 188: Artificial Intelligence Spring Recap Search II. Example State Space Graph. Search Problems.

Search and Optimization

Artificial Intelligence

CIS 192: Artificial Intelligence. Search and Constraint Satisfaction Alex Frias Nov. 30 th

Nearest neighbor classification DSE 220

Random Forest A. Fornaser

Foundations of Artificial Intelligence

Network Traffic Measurements and Analysis

6.034 Quiz 1, Spring 2004 Solutions

Chapter 6: Cluster Analysis

Informed search algorithms. (Based on slides by Oren Etzioni, Stuart Russell)

DFS. Depth-limited Search

Transcription:

Midterm Exam Thursday, March 15, 7:30 9:30 p.m. room 125 Ag Hall Covers topics through Decision Trees and Random Forests (does not include constraint satisfaction) Closed book 8.5 x 11 sheet with notes on both sides (typed or handwritten) Calculator (not on a phone) Covers lecture notes, readings in textbook (except Chapters 1 and 2), and 1 paper (intro to ML) True/False, multiple choice, and short answer questions Uninformed Search Methods Problem solving as search, problem representation in terms of states, goal test, operators, state-space graph search formulation, closed world assumption, Frontier and Explored lists, expanding a node, partial solution path, solution path, search tree Properties and computations completeness, optimality, time and space complexity breadth-first search depth-first search uniform-cost search iterative-deepening search bidirectional search Informed Search Methods Heuristic functions, evaluation function, admissible heuristic (h h*), consistent heuristic (h(n) c(n, n ) + h(n )), better informed heuristic, devising heuristics, completeness, admissibility best-first search greedy best-first search beam search algorithm A algorithm A* IDA* 1

Local Search Methods Local search problem formulation, operators, neighborhood (aka move set), local optima problem : hill-climbing algorithm hill-climbing with random restarts, stochastic hill-climbing (simulated annealing) escaping local optima, Boltzman s equation (p = e E / T ), temperature, cooling schedule genetic algorithms Computing crossover, mutation, fitness function, proportional fitness selection, next generation s population, crowding Genetic Algorithm (1 version*) 1. Let s = {s 1,, s N } be the current population 2. Let p[i] = f(s i )/SUM j f(s j ) be the fitness probabilities 3. for k = 1; k < N; k += 2 Parent1 = randomly pick s i with prob. p[i] Parent2 = randomly pick another s j with prob. p[j] Randomly select 1 crossover point, and swap strings of parents 1 and 2 to generate two children t[k] and t[k+1] 4. for k = 1; k N; k++ Randomly mutate each position in t[k] with a small probability 5. New generation replaces old generation: s = t *different than in book Game Playing Zero-sum games, perfect information games, deterministic vs. stochastic games, game playing as search, search tree, branching factor, ply, static evaluation function, horizon effect Minimax algorithm Minimax principle, optimal playing strategy Alpha-beta pruning algorithm Cutoff tests: If v α for some MAX node ancestor, don t visit any more of the current MIN node s children. If v β for some MIN node ancestor, don t visit any more of the current MAX node s children iterative-deepening search with alpha-beta pruning non-deterministic games chance nodes, expectimax value Monte Carlo tree search Pure MCTS, Selection, expansion, simulation, backpropagation, playout, exploitation, exploration MCTS Algorithm Recursively build search tree, where each iteration consists of: 1. Selection: Starting at root, successively select best child nodes using scoring method until leaf node L reached 2. Expansion: Create and add best (or random) new child node, C, of L 3. Simulation: Perform a (random) playout from C 4. Backpropagation: Update score at C and all of C s ancestors in search tree based on playout results 2

Unsupervised Learning Inductive learning problem, feature space, feature, attribute, examples (aka instances), labels, classes, training set, tuning set, testing set, classification problems, decision boundaries hierarchical agglomerative clustering single linkage, complete linkage, average linkage, Euclidean distance, Manhattan distance, dendrogram k-means clustering cluster center, distortion cluster quality measure: For all clusters, sum of squared distances from each point to its cluster center Nothing on mean-shift clustering Hierarchical Agglomerative Clustering Algorithm K-Means Algorithm Input: x 1,, x n, k where each x i is a point in a d- dimensional feature space Step 1: Select k cluster centers, c 1,, c k Step 2: For each point x i, determine its cluster: Find the closest center (using, say, Euclidean distance) Step 3: Update all cluster centers as the centroids Repeat steps 2 and 3 until cluster centers no longer change Supervised Learning Inductive bias, preference bias, decision boundaries, training set, tuning set, testing set, overfitting, noisy data, setting parameters K-nearest neighbor (k-nn) Decision Trees Ockham s razor, entropy, conditional entropy (aka remainder), information gain, categorical attributes, real-valued attributes, rule extraction, methods to avoid overfitting, pruning algorithm Random Forests Ensemble methods, bagging, bootstrap sampling, randomized node optimization, majority/mode classifier Performance evaluation Training set accuracy, training set error, testing set accuracy, testing set error, k-fold cross validation, leave-one-out cross validation 3

k-nearest-neighbors (k-nn) Entropy: Conditional Entropy: Decision boundary 1-NN: Information Gain: Decision-Tree-Learning Algorithm Pruning using a Greedy Algorithm buildtree(examples, attributes, default-label) if empty(examples) then return default-label if (examples all have same label y) then return y if empty(attributes) then return majority-class of examples q = best_attribute(examples, attributes) tree = create-node with attribute q foreach value v of attribute q do v-ex = subset of examples with q == v subtree = buildtree(v-ex, attributes - {q}, majority-class(examples)) add arc from tree to subtree return tree Prune(tree T, TUNE set) 1. Compute T s accuracy on TUNE; call it Acc(T) 2. For every internal node N in T: a) New tree T N = copy of T, but prune (delete) the subtree under N b) N becomes a leaf node in T N. The class is the majority vote of TRAIN examples reaching N c) Acc(T N ) = T N s accuracy on TUNE 3. Let T* be the tree (among the T N s and T) with the largest Acc() Set T = T * /* prune */ 4. If no improvement then Return T else Goto Step 1 4

Random Forests K-Fold Cross Validation For each tree, 1. Build a training set by choosing n times with replacement from all N available training examples (aka taking a bootstrap sample ) 2. At each node of decision tree during construction, choose a random subset of m attributes from the total number, M, of possible attributes (m << M) 3. Select the best attribute at node using Max-Gain 1. Divide all examples into K disjoint subsets E = E 1, E 2,..., E K 2. For each i = 1,..., K let TEST set = E i and TRAIN set = E - E i build decision tree using TRAIN set determine accuracy Acc i using TEST set 3. Compute K-fold cross-validation estimate of performance = mean accuracy = (Acc 1 + Acc 2 +... + Acc K )/K Leave-One-Out Cross Validation For i = 1 to N do // N = number of examples 1. Let (x i, y i ) be the i th example 2. Remove (x i, y i ) from the dataset 3. Train on the remaining N-1 examples 4. Compute accuracy on i th example Accuracy = mean accuracy on all N runs 5