Tracking system. Danica Kragic. Object Recognition & Model Based Tracking

Similar documents
Local Features Tutorial: Nov. 8, 04

Announcements. Recognition I. Gradient Space (p,q) What is the reflectance map?

Correspondence. CS 468 Geometry Processing Algorithms. Maks Ovsjanikov

Visual Learning and Recognition of 3D Objects from Appearance

Object Recognition with Invariant Features

Patch-based Object Recognition. Basic Idea

The SIFT (Scale Invariant Feature

3D Environment Reconstruction

Robot Mapping. Least Squares Approach to SLAM. Cyrill Stachniss

Graphbased. Kalman filter. Particle filter. Three Main SLAM Paradigms. Robot Mapping. Least Squares Approach to SLAM. Least Squares in General

Classifying Images with Visual/Textual Cues. By Steven Kappes and Yan Cao

Building a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882

Automatic Image Alignment (feature-based)

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

Generalized Principal Component Analysis CVPR 2007

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

Clustering Color/Intensity. Group together pixels of similar color/intensity.

Topological Mapping. Discrete Bayes Filter

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Eppur si muove ( And yet it moves )

TA Section 7 Problem Set 3. SIFT (Lowe 2004) Shape Context (Belongie et al. 2002) Voxel Coloring (Seitz and Dyer 1999)

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Capturing, Modeling, Rendering 3D Structures

Face Re-Lighting from a Single Image under Harsh Lighting Conditions

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier

Introduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.

Re-rendering from a Dense/Sparse Set of Images

Previously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011

Miniature faking. In close-up photo, the depth of field is limited.

A Statistical Approach to Culture Colors Distribution in Video Sensors Angela D Angelo, Jean-Luc Dugelay

Other approaches to obtaining 3D structure

Photo by Carl Warner

Outline 7/2/201011/6/

Computer vision: models, learning and inference. Chapter 13 Image preprocessing and feature extraction

Computer Vision for HCI. Topics of This Lecture

Stereo Wrap + Motion. Computer Vision I. CSE252A Lecture 17

Dimension Reduction CS534

Energy Minimization for Segmentation in Computer Vision

Contents I IMAGE FORMATION 1

Feature Detectors and Descriptors: Corners, Lines, etc.

Scale Invariant Feature Transform by David Lowe

CS 534: Computer Vision 3D Model-based recognition

Surface Registration. Gianpaolo Palma

3D Perception. CS 4495 Computer Vision K. Hawkins. CS 4495 Computer Vision. 3D Perception. Kelsey Hawkins Robotics

Feature descriptors and matching

INFO0948 Fitting and Shape Matching

CS4670: Computer Vision

Key properties of local features

Multiple-Choice Questionnaire Group C

COMP 175 COMPUTER GRAPHICS. Ray Casting. COMP 175: Computer Graphics April 26, Erik Anderson 09 Ray Casting

CS231A Midterm Review. Friday 5/6/2016

Local Feature Detectors

Local features: detection and description. Local invariant features

CS231A Section 6: Problem Set 3

Case-Based Reasoning. CS 188: Artificial Intelligence Fall Nearest-Neighbor Classification. Parametric / Non-parametric.

CS 188: Artificial Intelligence Fall 2008

Last update: May 4, Vision. CMSC 421: Chapter 24. CMSC 421: Chapter 24 1

Geometric Registration for Deformable Shapes 3.3 Advanced Global Matching

Computer Vision 6 Segmentation by Fitting

Motion illusion, rotating snakes

Scale Invariant Feature Transform

HOUGH TRANSFORM CS 6350 C V

3-D Object Recognition Using 2-D Views Wenjing Li, Member, IEEE, George Bebis, Member, IEEE, and Nikolaos G. Bourbakis, Fellow, IEEE

Multiview Stereo COSC450. Lecture 8

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Nonparametric Clustering of High Dimensional Data

Recognition: Face Recognition. Linda Shapiro EE/CSE 576

/10/$ IEEE 4048

CITS 4402 Computer Vision

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Local features: detection and description May 12 th, 2015

Digital Image Processing COSC 6380/4393

Beyond Bags of features Spatial information & Shape models

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

EE795: Computer Vision and Intelligent Systems

3D Object Recognition using Multiclass SVM-KNN

Scale Invariant Feature Transform

CV: Matching in 2D. Matching 2D images to 2D images; Matching 2D images to 2D maps or 2D models; Matching 2D maps to 2D maps MSU CSE 803

3D Models and Matching

Advanced Techniques for Mobile Robotics Graph-based SLAM using Least Squares. Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz

Data fusion and multi-cue data matching using diffusion maps

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

Camera calibration. Robotic vision. Ville Kyrki

Geometric Registration for Deformable Shapes 2.2 Deformable Registration

Shape Modeling with Point-Sampled Geometry

Shape Context Matching For Efficient OCR

A Novel Algorithm for Color Image matching using Wavelet-SIFT

Digital Geometry Processing

Final Exam Study Guide CSE/EE 486 Fall 2007

Figure (5) Kohonen Self-Organized Map

Extracting Layers and Recognizing Features for Automatic Map Understanding. Yao-Yi Chiang

Task analysis based on observing hands and objects by vision

Lecture 19: Motion. Effect of window size 11/20/2007. Sources of error in correspondences. Review Problem set 3. Tuesday, Nov 20

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

Dynamic Geometry Processing

The Pre-Image Problem in Kernel Methods

EXAM SOLUTIONS. Computer Vision Course 2D1420 Thursday, 11 th of march 2003,

DD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication

Transcription:

Tracking system Object Recognition & Model Based Tracking

Motivation Manipulating objects in domestic environments Localization / Navigation Object Recognition Servoing Tracking Grasping Pose estimation

Recognition (2D) Tracking (2D) Pose estimation (3D) : Initial pose estimation Steps Where in the image? Where in the world?

Initial Pose Estimation Recognition/Tracking Pose estimation (x,y) (X,Y,Z, φ, ψ, γ)

Example Objects Object Recognition & Model Based Tracking

Characteristics Simple geometry (polyhedra, cones, cylinders) Specular surfaces Background Illumination Slippery objects

Characteristics Simple geometry wireframe models Specular surfaces - ll Illumination - ll Background - ll Highly texture appearance Slippery objects power grasps

Model Based Techniques Appearance based methods Geometry based methods 3D wireframe models Complete pose estimation FUSION! Techniques from computer graphics used for rendering

Removes background, preserves object. Necessary to raise the signal to noise ratio, for the pose estimatior. Solved using color cooccurrence histograms. Object Recognition

An apperance based method is used to recognize the object, and estimate an initial pose. A geometric model based method is used to obtain an accurate pose. Algorithm combines the robustness of appearance based methods with the accuracy of feature based methods. Pose Estimation

Object Recognition & Model Based Tracking

Object Recognition & Model Based Tracking

Color Cooccurrence Histograms Apperance based method. Based on color cues only. Superior to standard color histograms. Invariant to translation and rotation. Robust towards scale changes.

Building Color Cooccurrence Histograms All pairs of pixels within a certain radius contribute to the histogram. Example: 4x4 image with 3 colors, and a maximum radius of 3 pixels. Histogram:

Building Color Cooccurrence Histograms When all pairs have been counted, the histogram is normalized. Each bin is divided with the total number of pixel pairs. 50 % Histogram:

Color Cooccurrence Histograms - Matching A common histogram matching method is used. Match = N i= 1 min( h1 [ i], h2[ i]) Reduces the effect of background noise, as unexpected colors will not penalize the match value.

Before the histogram can be built, the colors in the image need to be quantized. This is done using k- means clustering. Color Quantization Red Green

Images are normalized prior to quantization, in order to decrease the effect of varying lighting conditions. Only the red and green components are preserved. Performance equal to RGB and HSV. r r = norm r + g + b g g = norm r + g + b Color Quantization Red Green

Color Constancy Problem If lighting conditions change, colors may fall out of their original cluster, or even worse, into another one. Red green light Green

Object Segmentation - The system was trained using both front and back sides of the objects. The background of the training images was manually removed before training. Training

Object Segmentation A search window scans through the image, comparing the cooccurrence histogram with the stored histogram from the training images. The result is a vote matrix.

Object Segmentation From the vote matrix, segmentation windows are contructed. Starting from the global maximum, adjacent rows and columns are added as long as the vote values give sufficient support.

Object Segmentation - Out of 50 test images, 49 objects were successfully segmented. Average segmentation time was 1.7 s on a 500 MHz Sun station. Results

Pose Estimation The geometric model based pose estimator requires an initial pose to converge. The initial pose is estimated using color cooccurrence histograms.

Pose Estimation - Training 70 training images were used. The pose of the object varied over the training images. The correct pose of the object in the training image was stored, together with the cooccurrence histogram.

The object with the unknown pose is compared to each of the training examples. The result is a match value graph. Pose Estimation

The match value graph is filtered using a Gaussian kernel. Superior method compared to a nearest-neighbor approach. Pose Estimation

Initial Pose Estimation Appearance based Object Recognition & Model Based Tracking

Principle Component Analisys Learning stage compressing image set using eigenspace representation PCA PCA Pose recognition stage closest point search on appearance manifold PCA Fitting stage closest line search for pose refinement

PCA Pose Appearance Eigenstructure decomposition problem PCA i(q)

Implicit covariance matrix (conjugate gradient method) PCA PCA

Pose determination PCA PCA Object Recognition & Model Based Tracking

Initialization by PCA Object Recognition & Model Based Tracking

Geometric Model Based Pose Estimation Finally, the algorithm was integrated with the model based pose estimator.

Geometric Model Based Pose Estimation

Local refinement by tracking H = (14 0 60 15 6 5) [mm, deg]

Modeling Object Recognition & Model Based Tracking

Modeling Object Recognition & Model Based Tracking

Pose estimation DeMenthon and Davis 1995 Orthographic projection Iterative method No initial guess needed This step is followed by an extension of Lowe s nonlinear approach (Canceroni, Araujo and Brown et al.)

Lie algebra approach Tracking Rigid body motion SE(3) (6D Lie group) G G ω t x x = = 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 G t G y ω = y = 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 0 G t z = G ω 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 z = 0 1 0 0 1 0 0 0 0 0 0 0 0 0 0 0

with: Image motion = AE and x = u w vu Z y / / 1 Y X v w w + = 2 = AEG x = w w i Z and Li y v vw + 1 2 L - observed motion in an image point i w vu Y X u w uw w

Normal flow Object Recognition & Model Based Tracking

Rendering example Object Recognition & Model Based Tracking

3D pose update The change in pose is estimated using least square approach: 1 α = Cij O i i where α i represents the quantities of Euclidian motion O C i ij = = p p d p ( L ( L p i n p i p n p ) )( L p j n p )

3D pose update M I + α i G i t + 1 H ( R, t ) = H ( R, t ) ( I + α t i G i ) Object Recognition & Model Based Tracking

Examples Object Recognition & Model Based Tracking

Examples Object Recognition & Model Based Tracking

Example Object Recognition & Model Based Tracking

Task 1 Align and Track Object Recognition & Model Based Tracking

Task 1 Align and Track Object Recognition & Model Based Tracking

Task 2 Object Positioning Object Recognition & Model Based Tracking

Object Recognition & Model Based Tracking

Object Recognition & Model Based Tracking

Task 3 - Insertion Object Recognition & Model Based Tracking

How much a-priori info can we used? Insertion task Object Recognition & Model Based Tracking

Pick and Place Object Recognition & Model Based Tracking