Arnold W.M Smeulders Theo Gevers. University of Amsterdam smeulders}

Similar documents
HOW USEFUL ARE COLOUR INVARIANTS FOR IMAGE RETRIEVAL?

Deep Learning in Image and Face Intrinsics

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

HISTOGRAMS OF ORIENTATIO N GRADIENTS

Color Invariant Snakes

Chamfer matching. More on template matching. Distance transform example. Computing the distance transform. Shape based matching.

What is Computer Vision? Introduction. We all make mistakes. Why is this hard? What was happening. What do you see? Intro Computer Vision

Image Features: Detection, Description, and Matching and their Applications

Color. making some recognition problems easy. is 400nm (blue) to 700 nm (red) more; ex. X-rays, infrared, radio waves. n Used heavily in human vision

Physics-based Vision: an Introduction

Local Feature Detectors

Reflectance-based Classification of Color Edges

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Image Segmentation and Similarity of Color-Texture Objects

Announcements. Recognition I. Gradient Space (p,q) What is the reflectance map?

TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK

Last update: May 4, Vision. CMSC 421: Chapter 24. CMSC 421: Chapter 24 1

Color and Shading. Color. Shapiro and Stockman, Chapter 6. Color and Machine Vision. Color and Perception

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

TEXTURE CLASSIFICATION BY LOCAL SPATIAL PATTERN MAPPING BASED ON COMPLEX NETWORK MODEL. Srisupang Thewsuwan and Keiichi Horio

Yudistira Pictures; Universitas Brawijaya

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

CEng 477 Introduction to Computer Graphics Fall 2007

Multiple-Choice Questionnaire Group C

Different types of images. Taxonomy of fixed images. Introduction to image databases. Chapter VI: Image Databases

So far, we have considered only local models of illumination; they only account for incident light coming directly from the light sources.

Aalborg Universitet. A new approach for detecting local features Nguyen, Phuong Giang; Andersen, Hans Jørgen

Computer Vision I - Appearance-based Matching and Projective Geometry

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Evaluation and comparison of interest points/regions

Robotics Programming Laboratory

Visualisatie BMT. Rendering. Arjan Kok

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

Automatic Image Alignment (feature-based)

Local features: detection and description. Local invariant features

Evaluating Color and Shape Invariant Image Indexing of Consumer Photography T. Gevers and A.W.M. Smeulders Faculty of Mathematics & Computer Science,

Moving cast shadow detection of vehicle using combined color models

2 Related Work Very large digital image archives have been created and used in a number of applications including archives of images of postal stamps,

Visualization Concepts

Shadows in the graphics pipeline

Digital Vision Face recognition

Color Constancy from Hyper-Spectral Data

Outline 7/2/201011/6/

Colour Reading: Chapter 6. Black body radiators

[2006] IEEE. Reprinted, with permission, from [Wenjing Jia, Huaifeng Zhang, Xiangjian He, and Qiang Wu, A Comparison on Histogram Based Image

Object color forma9on

Digital Image Processing

CS 664 Segmentation. Daniel Huttenlocher

Chapter 4. Chapter 4. Computer Graphics 2006/2007 Chapter 4. Introduction to 3D 1

Computational Photography and Video: Intrinsic Images. Prof. Marc Pollefeys Dr. Gabriel Brostow

Announcements. Stereo Vision Wrapup & Intro Recognition

Ray-Tracing. Misha Kazhdan

Local Illumination. CMPT 361 Introduction to Computer Graphics Torsten Möller. Machiraju/Zhang/Möller

Local Image preprocessing (cont d)

An Algorithm to Determine the Chromaticity Under Non-uniform Illuminant

12/3/2009. What is Computer Vision? Applications. Application: Assisted driving Pedestrian and car detection. Application: Improving online search

General Principles of 3D Image Analysis

Computer Vision I - Appearance-based Matching and Projective Geometry

Object Recognition with Invariant Features

TA Section 7 Problem Set 3. SIFT (Lowe 2004) Shape Context (Belongie et al. 2002) Voxel Coloring (Seitz and Dyer 1999)

3D Modeling using multiple images Exam January 2008

Retrieving images based on a specific place in a living room

Light source estimation using feature points from specular highlights and cast shadows

Basic distinctions. Definitions. Epstein (1965) familiar size experiment. Distance, depth, and 3D shape cues. Distance, depth, and 3D shape cues

Illumination and Reflectance

Classification and Detection in Images. D.A. Forsyth

Object detection using Region Proposals (RCNN) Ernest Cheung COMP Presentation

SCALE INVARIANT FEATURE TRANSFORM (SIFT)

CS 534: Computer Vision Segmentation and Perceptual Grouping

Game Programming. Bing-Yu Chen National Taiwan University

CHAPTER 3. Single-view Geometry. 1. Consequences of Projection

An Introduction to Content Based Image Retrieval

Illumination and Shading

Implementation of 3D Object Recognition Based on Correspondence Grouping Final Report

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Loop detection and extended target tracking using laser data

Announcements. Recognition. Recognition. Recognition. Recognition. Homework 3 is due May 18, 11:59 PM Reading: Computer Vision I CSE 152 Lecture 14

2D Image Processing Feature Descriptors

Pipeline Operations. CS 4620 Lecture 10

Color Constancy from Illumination Changes

Segmentation by Clustering

Computer Vision for HCI. Topics of This Lecture

Visual Learning and Recognition of 3D Objects from Appearance

Supervised Sementation: Pixel Classification

Object and Class Recognition I:

ANDROID BASED OBJECT RECOGNITION AND MOTION DETECTION TO AID VISUALLY IMPAIRED

Simultaneous surface texture classification and illumination tilt angle prediction

AUTOMATIC IDENTIFICATION OF PORTRAITS IN ART IMAGES DATABASES

Image processing and features

ELEC Dr Reji Mathew Electrical Engineering UNSW

Photo-realistic Renderings for Machines Seong-heum Kim

Lecture 15: Shading-I. CITS3003 Graphics & Animation

Motion Tracking and Event Understanding in Video Sequences

Announcements. Introduction. Why is this hard? What is Computer Vision? We all make mistakes. What do you see? Class Web Page is up:

Computer Vision. Image Segmentation. 10. Segmentation. Computer Engineering, Sejong University. Dongil Han

CS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing

Digital Image Fundamentals

Local Features Tutorial: Nov. 8, 04

Pipeline Operations. CS 4620 Lecture Steve Marschner. Cornell CS4620 Spring 2018 Lecture 11

Journal of Asian Scientific Research FEATURES COMPOSITION FOR PROFICIENT AND REAL TIME RETRIEVAL IN CBIR SYSTEM. Tohid Sedghi

Transcription:

Arnold W.M Smeulders Theo evers University of Amsterdam email: smeulders@wins.uva.nl http://carol.wins.uva.nl/~{gevers smeulders}

0 Prolem statement Query matching Query

0 Prolem statement Query classes 1. Query y ey-words 2. Query y feature attriutes including setch 3. Query y 1 or N examples 4. Feedac iterative definition of query

1 Demands World model WWW Trademar assessment Trademar in pulic Stolen goods retrieval Video dataases Product dataase Product retrieval Stamp collections general pictures unnown conditions 2D of 2D pictures nown camera 2D of 3D pictures unnown conditions general pictures well-ehaved camera general pictures unnown camera limited domain nown camera limited domain well-ehaved camera 2D of 2D pictures narrow domains

1 Demands Computational model geometric models detect cue group line detect line controller data detect edge

1 Demands Computational model symolic models oject_one_end x = house U1 = 0.8 oject_other_end x = gasline U2 = 0.7 x = house to gasline Ui = 0.7 ojectz = house to gasline Ui = 0.7 z = douly arrowed line Uii = 0.63.

1 Demands World and computational models conclusions 1. Visual worlds have very specific characteristics ojects scenes sensors 2. The world more models than detailed computational can e made. 3.a Narrow world domains requires detailed computational models. 3. road domain require general models ased on physical laws light sensing and materials physiological facts perception cultural oservations semiotics.

1 Demands Image segmentation reconsidering the definition Input : picture data Output : find what pixels form an oject in real world Segmentation requires experience with meaning!

1 Demands Image segmentation advanced Segmentation y elastic snae model Contour which minimises elastic energy v s 2 fvs Advanced segmentation suited for narrow domain.

1 Demands Image segmentation conclusions 1. Whatever rilliant techniques for segmentation: not good enough to segment general pictures. 2. Segmentation has no answer to occlusion and clutter. 3. For retrieval of general images: wea segmentation: esult is definitely in one oject no guarantee all of oject is found. Simplest form: point segmentation.

1 Demands Features one million views So many images of one oject due to minor differences in: camera location rotation scale light source illumination light interaction surface cover acground clutter occlusion camera type viewpoint A million different data arrays to one oject!

1 Demands Features invariance Image retrieval in general domains requires invariance! A feature F invariant to condition A will produce for oject x the value Fx regardless the effect of condition A. Features for general image retrieval condition A effect of A required feature everywhere any positions location invariance upside down any orientation rotation invariance closer any scale scale invariance shadow lamp any intensity illumination invariance occlusion any suset present invariance to partial no frontal view any view angle viewpoint invariance

1 Demands Features invariance Image retrieval requires invariance! A feature F invariant to condition A will produce for oject x the value Fx regardless the effect of condition A. Image retrieval requires selected invariance! Art = frontal view daylight no occlusion. Crowd = occluded aritrary viewpoint.

1 Demands Features the invariance of 42 the numer Image retrieval requires invariance! A feature F invariant to condition A will produce for oject x the value Fx regardless the effect of condition A. The most invariant feature is the value 42 or any other numer: never changes regardless whatever conditions A. Trade-off etween invariance and power to discriminate. Conclusion: Use the smallest set of relevant invariant features.

1 Demands Features to recognize an oject One or two details may e enough to recognise an oject if and only if recognition is invariant colour changes shape changes

1 Demands Features conclusions 1. Invariant to sensing conditions illumination and oject shape require photometric features 2D image of 3D world require affine invariant features 2. Invariant to emedding scene occlusion and clutter require local features 3. Use the smallest set of relevant invariant features. 4. Candidate features are shape colour & texture iff invariant for domain recording circumstances and query at hand

2 Colour shape What maes an image? Spectrum of source Surface reflectance eometry of oject Emedding in scene Spectrum of sensor and nothing else

2 Colour shape What maes an image? -space ow will ojects reflect in -space?

2 Colour shape What maes an image? Schafer s model ody surface

2 Colour shape What maes an image? Schafer s model C = m n s f λ e λ c λ dλ m n s v f λ e λ c λ d C s C s λ λ λ λ c eλ n s v f C λ surface aledo illumination oject surface normal illumination direction viewer s direction sensor sensitivity scene & viewpoint invariant scene dependent viewpoint variant scene dependent viewpoint variant scene dependent

cos for example the geometricterm is Lamertian i.e. Consider the ody reflection term: s n s n m d c e f s n m C C = = λ λ λ λ λ What maes an image? ody reflectance in ody reflectance in -space space 2 Colour shape

2 Colour shape What maes an image? ody reflectance in -space ow does -histogram loo loo lie?

2 Colour shape What maes an image? ody reflectance in -space

2 Colour shape What maes an image? surface reflectance in -space Consider the surface reflection term : m n s v s λ f C λ e λ c λ dλ where the geometric term is for example α only depends on n s and v. s the phong model cos n α where

2 Colour shape What maes an image? reflectance under white light reconsidered S Dichromatic reflection under white light CW = em n s c ems n s v c s f

2 Colour shape Colour invariance S I - space S ue: = arctan 3

C s n em C = For giving the red green and lue sensor response under white light. Further with } { C = λ λ λ d c f C C = Consider the ody reflection term: s n em s n em g = = s n em s n em = = s n em s n em r = = Colour invariance photometric invariance photometric invariance 2 Colour shape

Colour invariance photometric invariance 2 Colour shape. 4 3 4 3 where Then 2 2 2 2 2 2 2 2 2 2 etc...} e.g. {...} e.g. { u t s r q p C s n em s n em s n em s n em s n em s n em u t s r q p = = = = = = = =

2 Colour shape Colour invariance c = 1 photometric invariance - c1 c2 c3 space arctan max{ } c = 2 c 3 = arctan arctan max{ } max{ } em n s c 1 arctan arctan = = max{ em n s em n s } max{ }

. 3 3 where Then 2 2 2 etc...} e.g. {...} e.g. { u t s r q p C u t s r q p = = Colour invariance shiny ojects shiny ojects - l1 l2 l3 space l1 l2 l3 space 2 Colour shape

2 Colour shape Colour invariance m1 m2 m3 - space coloured light m 1 = 2 3 / 3 2 m 2 = 3 1 / 1 3 m 3 = 1 2 / 2 1

2 Colour shape Colour invariance conclusion shadows shading highlights ill. intensity ill. color I - - - - - - - - - - rg - - c1c2c3 - - hue - l1l2l3 - m1m2m3 - - no invariance invariance

2 Colour shape Shape invariant colour edge detection c1c2c3 l1l2l3

2 Colour shape Shape colour edge classification color edge maxima shadows and geometry highlights color edges

2 Colour shape Shape colour edge classification material highlight shadow or geometry

3 Searching and finding Searching individual images

3 Searching and finding Searching individual images inary VSM - results 100 80 xor and Percent 60 40 mrs hausdorf 20 Accumulated raning percentile 0 1 2 3 4 5 10

3 Searching and finding Searching individual images inary VSM - roust against occlusion? accumulated 100% average ran 50% Yes! Even 80% occluded. 0 50 80 90 occlusion percentage

3 Searching and finding Searching individual images inary VSM - roust against viewpoints? accumulated 100% average ran 50% 0 45 60 7580 Yes! Even 75 o out of sight. rotation in viewpoint

3 Searching and finding Searching individual images conclusions on VSM 1. VSM wors fine for multi-coloured ojects shape colour wors est single est feature is colour. 2. VSM is roust against features: viewpoint affine invariant similarity: occlusion & clutter invariant

3 Searching and finding Searching groups of images VSM with -NN learning K-nearest neighor represents each images y a vector in n-space. The nearness is defined y the Euclidean distance. feature2 Training samples class A Training samples class Query image feature1

3 Searching and finding Searching groups of images VSM in -NN photographs versus synthetic synthetic images photographic images

3 Searching and finding Searching groups of images VSM in -NN photographs versus synthetic Color variation Synthetic images tend to have fewer colors than photographs. Color saturation Colors in synthetic images are commonly less saturated. Color edge strength Synthetic images tend to have more arupt color transitions than photographs.

3 Searching and finding Searching groups of images -NN in VSM classification results Query Query esults esults Edge strength and saturation photo synthetic Classifier y -NN Data 200 training and 100 images for each class. photo 95% 5% synthetic 9% 91%

3 Searching and finding Searching groups of images -NN in VSM for sin detection sin pixels not-sin pixels - space versus c1 c2 c3 - space Conclusion: a range in normalized c1 c2 c3-space specific for sin.

3 Searching and finding Searching groups of images sin detection: results

3 Searching and finding Searching groups of images K-NN in VSM for portrait detection Query esults portrait non-portrait Features are color ratio s. Classification y -NN. portrait 96% 4% Data 200 training and 100 test images in each class. non-portrait 25% 75%

4 Oject localisation Split and merge Split regions until patch is homogeneous...

4 Oject localisation Split and merge... and merge patches which are alie. Wors ecause of spatial coherence.

4 Oject localisation omogeneity 1. Normalizedcross correlation : D C IQ = D IQ I = = 1 t t t = 1 = 1 w Q w t = 1 w Q 2. istogram intersection : min{ w w I 2 Q Q w I } 10 0 0 D I and # D C I Q dependent on occlusion and oject cluttering I Q independen oject I cluttering Q hue t on occlusion

4 Oject localisation Data set example

4 Oject localisation esults

4 Oject localisation esults

5 Image search engine PicToSee Content-ased image retrieval Fast indexing Query pictorial example attriutes Invariance University of Amsterdam {gevers smeulders}@science.uva.nl http://carol.wins.uva.nl/~{gevers smeulders}