Recognition (Part 4) Introduction to Computer Vision CSE 152 Lecture 17

Similar documents
Announcements. Recognition (Part 3) Model-Based Vision. A Rough Recognition Spectrum. Pose consistency. Recognition by Hypothesize and Test

Recognition. Computer Vision I. CSE252A Lecture 20

CS 534: Computer Vision 3D Model-based recognition

Registration D.A. Forsyth, UIUC

Object Recognition 1

Object Recognition 1

Object and Class Recognition I:

Template Matching Rigid Motion

Template Matching Rigid Motion. Find transformation to align two images. Focus on geometric features

Announcements. Recognition I. Gradient Space (p,q) What is the reflectance map?

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Announcements. Stereo Vision Wrapup & Intro Recognition

Methods for Representing and Recognizing 3D objects

Instance-level recognition

High-Level Computer Vision

Using Geometric Blur for Point Correspondence

Instance-level recognition

Announcements. Recognition I. Optical Flow: Where do pixels move to? dy dt. I + y. I = x. di dt. dx dt. = t

Announcements. Recognition. Recognition. Recognition. Recognition. Homework 3 is due May 18, 11:59 PM Reading: Computer Vision I CSE 152 Lecture 14

Geometric Hashing. Overview (1) Overview (2) Reference Frames. Recognition under Similarity Transformations. Feb Goo Jun

Local features: detection and description May 12 th, 2015

Today. Introduction to recognition Alignment based approaches 11/4/2008

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Instance-level recognition II.

3D Object Modeling and Recognition Using Affine-Invariant Patches and Multi-View Spatial Constraints

Local features and image matching. Prof. Xin Yang HUST

Local features: detection and description. Local invariant features

Combining Appearance and Topology for Wide

Shape recognition with edge-based features

Instance-level recognition part 2

Patch-based Object Recognition. Basic Idea

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability

Render. Predicted Pose. Pose Estimator. Feature Position. Match Templates

Face detection by learned affine correspondences

Local Image Features. Synonyms. Definition. Introduction

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

Visual Object Recognition

Object Recognition in High Clutter Images Using Line Features

What is Computer Vision? Introduction. We all make mistakes. Why is this hard? What was happening. What do you see? Intro Computer Vision

N3M: Natural 3D Markers for Real-Time Object Detection and Pose Estimation

3D Models and Matching

Object Recognition with Invariant Features

Lecture 3 Sections 2.2, 4.4. Mon, Aug 31, 2009

Feature Matching and Robust Fitting

Auto-calibration. Computer Vision II CSE 252B

CS 534: Computer Vision Texture

Determining pose of a human face from a single monocular image

Multiple-Choice Questionnaire Group C

On Resolving Ambiguities in Arbitrary-Shape extraction by the Hough Transform

Detecting Geometric Primitives in 3D Data

Computer Vision I - Appearance-based Matching and Projective Geometry

Recognizing Deformable Shapes. Salvador Ruiz Correa Ph.D. UW EE

COMPUTER AND ROBOT VISION

Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images

Using Projective Invariants for Constant Time Library Indexing in Model Based Vision

Invariant Features from Interest Point Groups

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

CS4670: Computer Vision

Computer Vision I - Appearance-based Matching and Projective Geometry

CS231A Midterm Review. Friday 5/6/2016

Using Genetic Algorithms for Model-Based Object Recognition

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

A General Expression of the Fundamental Matrix for Both Perspective and Affine Cameras

Lucas-Kanade Scale Invariant Feature Transform for Uncontrolled Viewpoint Face Recognition

Today s lecture. Image Alignment and Stitching. Readings. Motion models

EE795: Computer Vision and Intelligent Systems

Recognition. Clark F. Olson. Cornell University. work on separate feature sets can be performed in

From Structure-from-Motion Point Clouds to Fast Location Recognition

CSE 252B: Computer Vision II

Data Representation in Visualisation

Large-Scale 3D Point Cloud Processing Tutorial 2013

Recognizing Deformable Shapes. Salvador Ruiz Correa (CSE/EE576 Computer Vision I)

Beyond Bags of features Spatial information & Shape models

Detecting Multiple Symmetries with Extended SIFT

CSE528 Computer Graphics: Theory, Algorithms, and Applications

3D Models and Matching

Face Detection and Recognition in an Image Sequence using Eigenedginess

Introduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.

Fuzzy based Multiple Dictionary Bag of Words for Image Classification

Prof. Feng Liu. Spring /26/2017

Announcements. Hough Transform [ Patented 1962 ] Generalized Hough Transform, line fitting. Assignment 2: Due today Midterm: Thursday, May 5 in class

Homogeneous Coordinates. Lecture18: Camera Models. Representation of Line and Point in 2D. Cross Product. Overall scaling is NOT important.

Topological Mapping. Discrete Bayes Filter

Object Recognition Using Pictorial Structures. Daniel Huttenlocher Computer Science Department. In This Talk. Object recognition in computer vision

Computer Vision Projective Geometry and Calibration. Pinhole cameras

Local Features Tutorial: Nov. 8, 04

Final Exam Study Guide

Recognizing Buildings in Urban Scene of Distant View ABSTRACT

Metric Rectification for Perspective Images of Planes

3D Polygon Rendering. Many applications use rendering of 3D polygons with direct illumination

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation

Planar Region Detection and Motion Recovery

CS664 Lecture #21: SIFT, object recognition, dynamic programming

Announcements. Introduction. Why is this hard? What is Computer Vision? We all make mistakes. What do you see? Class Web Page is up:

MULTI-VIEW FACE DETECTION AND POSE ESTIMATION EMPLOYING EDGE-BASED FEATURE VECTORS

Image Based Rendering. D.A. Forsyth, with slides from John Hart

Instance-level recognition I. - Camera geometry and image alignment

Gaze interaction (2): models and technologies

INFO0948 Fitting and Shape Matching

Global Gabor features for rotation invariant object classification

Transcription:

Recognition (Part 4) CSE 152 Lecture 17

Announcements Homework 5 is due June 9, 11:59 PM Reading: Chapter 15: Learning to Classify Chapter 16: Classifying Images Chapter 17: Detecting Objects in Images

A Rough Recognition Spectrum Appearance-Based Recognition (Eigenface, Fisherface) Shape Contexts Geometric Invariants Image Abstractions/ Volumetric Primitives Local Features + Spatial Relations Aspect Graphs 3-D Model-Based Recognition Function

Model-Based Vision Given 3-D models of each object Detect image features (often edges, line segments, conic sections) Establish correspondence between model & image features Estimate pose Consistency of projected model with image

Recognition by Hypothesize and Test General idea Hypothesize object identity and pose Recover camera parameters Render object using camera parameters Compare to image Issues Where do the hypotheses come from? How do we compare to image (verification)? Simplest approach Construct a correspondence for all object features to every correctly sized subset of image points These are the hypotheses Expensive search, which is also redundant

Pose consistency Correspondences between image features and model features are not independent A small number of correspondences yields a camera matrix The others correspondences must be consistent with this Strategy: Generate hypotheses using small numbers of correspondences (e.g., triples of points for a calibrated perspective camera) Recover camera parameters (e.g., calibrated camera rotation and translation) and verify

Example Figure from Object recognition using alignment, D.P. Huttenlocher and S. Ullman, Proc. Int. Conf. Computer Vision, 1986, copyright IEEE, 1986

Voting on Pose Each model leads to many correct sets of correspondences, each of which has the same pose Vote on pose, in an accumulator array (similar to a Hough transform)

Figure from The evolution and testing of a model-based object recognition system, J.L. Mundy and A. Heller, Proc. Int. Conf. Computer Vision, 1990 copyright 1990 IEEE

Figure from The evolution and testing of a model-based object recognition system, J.L. Mundy and A. Heller, Proc. Int. Conf. Computer Vision, 1990 copyright 1990 IEEE

Figure from The evolution and testing of a model-based object recognition system, J.L. Mundy and A. Heller, Proc. Int. Conf. Computer Vision, 1990 copyright 1990 IEEE

Figure from The evolution and testing of a model-based object recognition system, J.L. Mundy and A. Heller, Proc. Int. Conf. Computer Vision, 1990 copyright 1990 IEEE

Invariance Properties or measures that are independent of some group of transformation (e.g., rigid, affine, projective, etc.) For example, under affine transformations: Collinearity Parallelism Intersection Distance ratio along a line Angle ratios of three intersecting lines Affine coordinates

Invariance There are geometric properties that are invariant to camera transformations Easiest case: view a plane object in scaled orthography Assume we have three base points P i (i=1..3) on the object Any other point on the object can be written as P k P 1 ka P 2 P 1 kb P 3 P 1 Now image points are obtained by multiplying by a planar affine transformation, so p k AP k AP 1 ka P 2 P 1 kb P 3 P 1 p 1 ka p 2 p 1 kb p 3 p 1 P 2 P k P 1 P 3 P k+1

Geometric hashing Vote on identity and correspondence using invariants Take hypotheses with large enough votes Building a table: Take all triplets of points on model image to be base points P 1, P 2, and P 3 Take every fourth point and compute ka and kb Fill up a table, indexed by ka and kb, with The base points and fourth point that yielded ka and kb The object identity

Figure from Efficient model library access by projectively invariant indexing functions, by C.A. Rothwell et al., Proc. Computer Vision and Pattern Recognition, 1992, copyright 1992, IEEE

Verification Edge score Are there image edges near predicted object edges? Very unreliable; in texture, answer is usually yes Oriented edge score Are there image edges near predicted object edges with the right orientation? Better, but still hard to do well Texture For example, does the spanner have the same texture as the wood?

Application: Surgery To minimize damage by operation planning To reduce number of operations by planning surgery To remove only affected tissue Problem Ensure that the model with the operations planned on it and the information about the affected tissue lines up with the patient Display model information supervised on view of patient Big Issue: coordinate alignment, as above

MRI CTI NMI USI Reprinted from Image and Vision Computing, v. 13, N. Ayache, Medical computer vision, virtual reality and robotics, Page 296, copyright, (1995), with permission from Elsevier Science

Figures by kind permission of Eric Grimson; further information can be obtained from his web site http://www.ai.mit.edu/people/welg/welg.html.

Figures by kind permission of Eric Grimson; further information can be obtained from his web site http://www.ai.mit.edu/people/welg/welg.html.

Figures by kind permission of Eric Grimson; further information can be obtained from his web site http://www.ai.mit.edu/people/welg/welg.html.

A Rough Recognition Spectrum Appearance-Based Recognition (Eigenface, Fisherface) Shape Contexts Geometric Invariants Image Abstractions/ Volumetric Primitives Local Features + Spatial Relations Aspect Graphs 3-D Model-Based Recognition Function

Matching using Local Image features Simple approach Detect corners in image (e.g., Harris corner detector) Represent neighborhood of corner by a feature vector (produced by Gabor Filters, K-jets, affine-invariant features, etc.) Modeling: Given a training image of an object without clutter, detect corners, and compute and store feature descriptors Recognition time: Given test image with possible clutter, detect corners and compute features. Find models with same feature descriptors (hashing) and vote

Figure from Local grayvalue invariants for image retrieval, by C. Schmid and R. Mohr, IEEE Trans. Pattern Analysis and Machine Intelligence, 1997 copyright 1997, IEEE

Write Probabilistic interpretation Assume Likelihood of image given pattern

Employ spatial relations Figure from Local grayvalue invariants for image retrieval, by C. Schmid and R. Mohr, IEEE Trans. Pattern Analysis and Machine Intelligence, 1997 copyright 1997, IEEE

Figure from Local grayvalue invariants for image retrieval, by C. Schmid and R. Mohr, IEEE Trans. Pattern Analysis and Machine Intelligence, 1997 copyright 1997, IEEE

Example Training examples Test image

Finding faces using relations Strategy: Face is eyes, nose, mouth, etc. with appropriate relations between them Build a specialized detector for each of these (template matching) and look for groups with the right internal structure Once a face is detected, there is little uncertainty about where the other parts could be

Finding faces using relations Strategy: compare Notice that once some facial features have been found, the position of the rest is quite strongly constrained. Figure from, Finding faces in cluttered scenes using random labelled graph matching, by Leung, T. ;Burl, M and Perona, P., Proc. Int. Conf. on Computer Vision, 1995 copyright 1995, IEEE

Figure from, Finding faces in cluttered scenes using random labelled graph matching, by Leung, T. ;Burl, M and Perona, P., Proc. Int. Conf. on Computer Vision, 1995 copyright 1995, IEEE

Color Reading: Chapter 3: Color Next Lecture