What does it mean to see?
|
|
- Doreen Lambert
- 5 years ago
- Views:
Transcription
1 What does it mean to see? The engineering science of computer vision James L. Crowley Professor, I.N.P. Grenoble Projet PRIMA - Laboratory GRAVIR INRIA Rhône Alpes Grenoble, France 1
2 The Science of Computer Vision Computer Vision is a branch of engineering science (Simon 69) that has developed through a series of paradigms. Science: The elaboration of theories and models that explain and predict. Science is a method of investigation performed by a collection of scientists (a scientific community) who share a set of paradigms. Paradigm: The problems and problem solutions adopted by a scientific community. T. S. Kuhn, The Structure of Scientific Revolutions, The Univ. of Chicago Press, Chicago, H. A. Simon, The Sciences of the Artificial, The MIT Press, Cambridge Mass.,
3 The Science of Computer Vision Computer vision is the science of machines that see. What does it mean for a "machine" to "see"? What are the paradigms of Computer Vision? 3
4 Paradigms for Computer Vision Early Paradigms Blocks World and Scene Analysis ( ) Symbolic Artificial Intelligence ( ) Established Paradigms 3D Reconstruction Active Vision Emerging paradigms Software Architectures for Vision Systems Appearance Based Vision Statistical Learning 4
5 What does it mean to see? Outline: The Science of Computer Vision Early Paradigms Geometric Foundations Active Vision Physics Based Vision: Color Models Real Time Tracking Systems Appearance Based Vision Possible Future Paradigms and Conclusions 5
6 What does it mean to see? Outline: The Science of Computer Vision Early Paradigms Geometric Foundations Active Vision Physics Based Vision: Color Models Real Time Tracking Systems Appearance Based Vision Possible Future Paradigms and Conclusions 6
7 The Blocks World Examples: The Thesis of L. Roberts (Stanford 1963) Pattern Recognition and Scene Analysis, (Duda and Hart, 1972) Line Labeling (Waltz, Huffman, etc) Approach: 1) Structural recognition of polyhedral objects 2) Edge Detection (Roberts, Sobel, etc) 3) Homogeneous Coordinates 4) Wire Frame 3-D Object Representation 7
8 The Blocks World Debates: Hypotheses: Failures: Edge Detection vs Segmentation Structural vs Syntactic vs Statistical Pattern Recognition The blocks world assumed static rigid planar surfaces with Lambertian albedo and diffuse illumination. The real world is composed of dynamic deformable non-planar objects with arbitrary albedo under arbitrary and varying illumination. Techniques from the blocks world were too fragile to be useful. 8
9 Symbolic Reasoning for Artificial Intelligence Dinner Table Contains Contains Plate Next To Problems: 1) Representing object knowledge 2) Representing knowledge of geometry and image formation 3) Inference Techniques Problem Solutions 1) Frames and Rules 2) Production Systems, Theorem Proving 3) Prediction and Verification (Context) fork 9
10 Symbolic Reasoning for Artificial Intelligence Dinner Table Contains Contains Debates: Plate Next To fork Top down or bottom up. Structural or symbolic models How to learn models Failure: Unstable image representations Computational Complexity Knowledge Acquisition 10
11 Symbolic Reasoning for Artificial Intelligence Dinner Table Contains Contains Plate fork examples: Frames (Minsky, 1975) Next To Interpretation Guided Segmentation (Barrow and Tennenbaum, 77) Visions (Hanson and Riseman, 1978) Rule based interpretation (Mckeown 1985) Schema System (Draper 87) 11
12 The Marr Paradigm D. Marr Vision, W. H. Freeman, San Francisco, (Plus other books by Grimson, Hildreth, Ullmann, etc) Concepts: Three layered architecture composed of Primal Sketch, 2 1/2 D Sketch, and Hierarchical Object Models. Inspiration from neuroscience. 12
13 The Marr Paradigm Problems: 1) Representing the Primal Sketch 2) Computing the 2 1/2 D Sketch 3) Interpolation 4) Object Model Representation Problem Solutions 1) Laplacian Pyramid 2) Zero Crossing Contours in the Laplacian of Gaussian 3) Shape from "X" 4) Generalized Cylinders 13
14 Vision as Reconstruction Debates: 1) Object or Camera centered reference frame? 2) Biological relevance of Computational Models 3) Fusion and Integration of Shape from "X". Marr insisted that the invariant in vision is the 3D world. This has led to 15 years of research on vision as exact 3-D reconstruction from multiple images. 14
15 What does it mean to see? Outline: The Science of Computer Vision Early Paradigms Geometric Foundations Active Vision Physics Based Vision: Color Models Real Time Tracking Systems Appearance Based Vision Possible Future Paradigms and Conclusions 15
16 Homogeneous Coordinates Image Point r P = x y 1 Image Line r L = a b c ( ) Scene Point r Q = x y z 1 Plane r S = a b c d ( ) Line Equation r L T r P = ax +by+c = 0 Plane Equation r S T r Q =ax +by+cz+d=0 16
17 Projective Camera Model Point pinhole Retina M s i = m 11 m 12 m 13 m 14 m 21 m 22 m 23 m 24 m 31 m 32 m 33 m 34 w r P i = M s i r Q s or wi wj w = x m 11 m 12 m 13 m 14 y m 21 m 22 m 23 m 24 z m 31 m 32 m 33 m
18 Homographic Projection Bijective projection from one plane to another a b H a b = h 11 h 12 h 13 h 21 h 22 h 23 h 31 h 32 h 33 w r P b = H a b r Q a or wx a wy a w = h 11 h 12 h 13 x b h 21 h 22 h 23 y b h 31 h 32 h
19 Fundamental Matrix R s Points in an image project to a line in a second image p a e a Base Line eb q b F AB = f 11 f 12 f 13 f 21 f 22 f 23 f 31 f 32 f 33 A s Image A r r L B = F AB Q A or image B B s a b = c f 11 f 12 f 13 x a f 21 f 22 f 23 y a f 31 f 32 f
20 Trifocal Tensor C s m c l c r R c = T ab c r P a r Q b e c r c Image C f c p a R s q b A s B s Image A Image B Geometric relation between points in three images Correspondence of 29 points in three images gives transformation for all other points 20
21 Multi-camera Geometry References : R. I. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision, Cambridge University Press, O. Faugeras, Three-dimensional Computer Vision: A Geometric Viewpoint, MIT Press,
22 What does it mean to see? Outline: The Science of Computer Vision Early Paradigms Geometric Foundations Active Vision Physics Based Vision: Color Models Real Time Tracking Systems Appearance Based Vision Possible Future Paradigms and Conclusions 22
23 Active Vision Approach: Use control of cameras to simplify observations of the external world. The observations are defined by the requirements of a task. An active vision system acts as a "FILTER" for information. Problems: 1) Integration : a) of processing modes b) with (robotics) applications 2) Robust, real time, image processing 3) Control of fixation and attention 4) Control of processing. 23
24 Active Vision Control of sensors and processing to make observations about the external world. Real Time: An active vision system must return its results within a fixed delay. Attention is limited: Fixed delay response requires limiting the data. This means restricting processing to a small region of interest. An active vision system acts as a FILTER for information. 24
25 Active Vision Continuous operation: The system is always running. The results at each instant provide the context for the next instant. Processing and parameters are determined "on the fly" The vision system is a series of filters : Fixation: The fixation point and the horopter Color: The color filters ROI: The region of interest Description: Receptive fields Attention: Context from task and activity 25
26 The LIFIA Camera Head (1991) 26
27 Platform for Visual Navigation Supervisor M a i l b o x Device Controllers VisionClips Processes Image Processing Silicon Graphics Fixation Camera Controller Navigation Sun Vehicle Controller Albatros 27
28 LIFIA Binocular Head 28
29 Why Fixate? 1) Fixation makes real time processing possible 2) Fixation cancels motion blur. 3) Fixation separates figure from ground. Challenge: Real time Control. Multi-cue Integration 29
30 The Human Eye Nerf Optique Fovéa (cônes) Périphéríe (bâtonnets) Fovéa Cornéa The retina is composed of the fovea and the periphery. Retina provides precise acuity for recognition. The periphery guides fixation and triggers reflexes. 30
31 The Horopter Horopter : The locus of points in the scene which map to (nearly) the same location in the two images. = +σ Point de Fixation = 0 The Horopter is a FILTER for information. = σ The Horopter permits a simple separation of figure and ground. B C D B C D 31
32 KTH Binocular Head (1992) 32
33 KTH Binocular Head (1992) 33
34 Vergence and Version: The Vief Muller Circle Zero Disparity Surface : h l = h r Fixation Point F P Interest point 2 µ Stereo Fusion Reflex: Map image points to the same image position Primary Line of Sight η l α l α r η r Corresponding Visual Rays (Much easier than Stereo Matching) Left Retina Right Retina Fixation Center 34
35 Vergence and Version: The Vief Muller Circle Z Z Z 2 η 2 η η η α l α r α l α r α l α r x α c x α c X Symmetric Vergence Vergence Version 35
36 KTH Binocular Head (1992) 36
37 Active Vision References: J.Y. Aloimonos, I. Weiss and A. Bandopadhay, "Active Vision", International Journal on Computer Vision, pp , R. Bajcsy, "Active Perception", IEEE Proceedings, Vol 76, No 8, pp , August Ballard, D.H. and Ozcandarli, A., Eye Fixation and Early Vision: Kinematic Depth, IEEE 2nd Intl. Conf. on Comp. Vision, Tarpon Springs, Fla., pp , Dec J. L. Crowley and H. I Christensen,Vision as Process, Springer Verlag,
38 What does it mean to see? Outline: The Science of Computer Vision Early Paradigms Geometric Foundations Active Vision Physics Based Vision: Color Models Real Time Tracking Systems Appearance Based Vision Possible Future Paradigms and Conclusions 39
39 Albedo: Reflectance Functions Reflectance Functions describe the interaction of light with matter. Lumiere R L (i, e, g, λ) = Number of photons emitted Number of photons received N i g Camèra e Lambertian Reflection (ex: Paper, snow) Specular Reflection (ex: Mirror) R L (i, λ)= P(λ)cos(i) R S (i, e, g, λ) = 1 if i = e and i +e=g 0 otherwise 40
40 Albedo: Reflectance Functions Arbitrary reflectance functions can be modeled as a weighted sum of Lambertian and Specular reflection. Composant Speculaire Composant Lambertian Lumieres R(i, e, g, λ) = c R S (i, e, g, λ) + (1-c) R L (i, λ) Surface Pigment Lambertian Reflection (ex: Paper, snow) Specular Reflection (ex: Mirror) R L (i, λ)= P(λ)cos(i) R S (i, e, g, λ) = 1 if i = e and i +e=g 0 otherwise 41
41 Luminance and Chrominance Lambertian reflection can be decomposed into luminance and chrominance Composant Speculaire Lumieres Lambertian Reflection: R L (i, λ) = P(λ)cos(i) Composant Lambertian Surface Pigment Luminance is determined by surface orientation (describes 3D shape) Chrominance identifies object pigments (signature for object recognition) 42
42 Color Perception Nerf Optique Fovéa (cônes) Périphéríe (bâtonnets) Fovéa Cornéa Day Vision: High Acuity, chromatic Three pigments: cyanolabe (445 nm), chlorolabe (535 nm), and erythrolabe (570 nm) Night Vision: Achromatic Single pigment - rhodopsine (510 nm) 43
43 Color Perception 1 Reponse Relative β γ 0.5 α nm λ Day Vision: High Acuity, chromatic Three pigments: cyanolabe (445 nm), chlorolabe (535 nm), and erythrolabe (570 nm) Color percetion is subjective, based on relative logarithmic responses of three color channels 44
44 Hue Luminance and Saturation HLS Color Space. Luminance: Relative Intensity (Black-white) Hue is angle on a plane Saturation is radial distance Plan Chromatique L Axe de Luminance S T 45
45 Color Perception 0.3 bleu λ b( ) vert rouge λ r( ) 0.2 g( λ ) nm λ Color cameras imitate RGB with three filters. RGB 46
46 Color Spaces: RGB magenta B bleu R rouge blanc noir cyan jaune vert Axe Achromatique Triangle de Maxwell [R + V + B = 1] V Plan Chromatique "complémentaire" [R + V + B = 2] RGB Color Cube Considers each filter as an orthogonal sample of the spectrum 47
47 Example : Color Skin detection Region of Interest Sample rate Color Table Color Image Skin Color Detection Average probability Execution time Skin Probability Transform RGB pixels in to probability of skin Theory: Bayes rule Implementation: table lookup 48
48 Probabilistic Detection of Skin Chrominance: r = R R + G + B g = G R + G + B Probability of all colors Probability of skin p(r,g) 1 N Tot h Tot (r,g) p(r,g skin) 1 N skin h skin (r,g) p(skin r,g) = p(r,g skin)p(skin) p(r,g) h skin(r,g) h Tot (r,g) = h ratio (r,g) 49
49 detecting skin pixels with color Sample Rate Color Table Average Probability Execution Time Skin Color Detector Using Chrominance to detect hands 50
50 What does it mean to see? Outline: The Science of Computer Vision Early Paradigms Geometric Foundations Active Vision Physics Based Vision: Color Models Real Time Tracking Systems Appearance Based Vision Possible Future Paradigms and Conclusions 51
51 Blob grouping ROI Blob Detected Probability Image Moment Based Grouping (Blob, ID, CF,x, y, s x, s y, θ) Blob: A connected region of detected pixels Properties: Position and spatial extent Theory: Moments 52
52 Blob grouping Confidence S = i max i= i min j max j= j min p skin (i,j) Position i max µ i = 1 S i=imin j max p skin (i,j) i i max µ j = 1 S i= i min j max p skin (i,j) j j= j min j= j min i max σ i 2 = 1 S i= imin j max p skin (i, j) (i µ i ) 2 j=j min Spatial Extent i max σ j 2 = 1 S i= imin i max σ ij = 1 S i= i min j max (i,j) (j µ j ) 2 j= j min j max p skin p skin (i, j) (i µ i ) (j µ j ) j= j min 53
53 54
54 Tracking: Recursive Estimation Y t,c y,cf Match X ˆ t, C ˆ t,cf Estimation Observation Update Detect Predict Region of Interest * X t + t *,C t + t,cf Tracking: 1) Optimise processing by focus of attention 2) Maintain target and ID properties across time. 55
55 Robust Tracking Multiply pixels with Gaussian function based on detection in previous image. f skin (i, j) := f skin (i, j) e 1 i µ i 2 j µj T i (kc) 1 µ i j µj p(skin t µ t- t, C t- t ) 56
56 57
57 Multi-Cue Face Tracking Supervisor Blink Detection Eye Detector (correlation) Face Detector (color) Blink Detection: Precise but infrequent Correlation: Fast and Precise but fragile Probabilistic Chrominance: Slower and less precise, but reliable. Approach: coordinate multipe redundant detection processes J. L. Crowley and F. Berard, "Multi-Modal Tracking of Faces for Video Communications", IEEE Conference on Computer Vision and Pattern Recognition, CVPR '97, St. Juan, Puerto Rico, June
58 Multi-Cue Face Tracking 59
59 Multi-Cue Face Tracking Supervisor Image Acquisition and Processing Blink Detection ColorDetection Correlation Tracking Camera Control Interpreter Blink SSD Color
60 Blue Eye Video Entity Detection and Tracking Process Corba Shell Time Detection Prediction Video Stream Video Demon ObservationMo ObservationMo dules Observation dules Modules Estimation Entities Event Detection Events Hardwired Control in C++ Communication using CORBA Observation Modules: Color Histogram Ratio, Background Difference, Motion History Image 61
61 PETS Benchmark #2 62
62 Blue Eye Video Activity Sensor (PETS 2002 Data) 63
63 Blue Eye Video Activity Sensor (Intersection Event Observation) 64
64 CAVIAR Indoor Test-bed: INRIA Entrance Hall 2 Cameras: one w/wide angle lens, one steerable pan-tilt-zoom 65
65 66
66 Left-behind Backage Detection 67
67 CAVIAR Outdoor Test Bed INRIA Back Parking Lot 2 Outdoor Surveilance Platforms, 3 m separation, 3 meter height 68
68 Back Parking Lot Behaviour Analysis 69
69 Tracking, Recognition and Attention Lesson: Detect, Track then Recognize. Tracking focusses attention for recognition Tracking: 1) Conserves identity 2) Focusses Processing Resources 3) Provides Robustness to noise 4) Permits Temporal Fusion 70
70 What does it mean to See? Outline: The Science of Computer Vision Early Paradigms Geometric Foundations Active Vision Real Time Tracking Systems Software Architectures Possible Future Paradigms Conclusions 71
71 Supervised Perceptual Process Events Configuration Requests for state Events Current State Response to commands Autonomic Supervisor Time ROI, S, Detection Method Detection Prediction ROI, S, Detection Method Video Stream ObservationMo ObservationMo dules Observation dules Modules Estimation Entities Intepretation Actors Supervisor Provides: Execution Scheduler Parameter Regulator Command Interpreter Description of State and Capabilities 72
72 Supervised Perceptual Process Process Phases: While True Do Acquire next image Calculate ROI for targets Verify and update targets Detect new targets Regulate module parameters Interpret entities Process messages 73
73 Autonomic Properties provided by process supervisor Auto-regulatory: The process controller can adapt parameters to maintain a desired process state. Auto-descriptive: The process controller provides descriptions of the capabilities and the current state of the process. Auto-critical: Process estimates confidence for all properties and events. 74
74 Categories of Processes Control in State and Capabilities Data Stream Control Entity detection and tracking Entities Events Entity Tracking Processes Input: Data: Sensor stream(video, Acoustic, tactile) Output: Data: List of entities with properties Events: Detection, Loss, Entry in Region, Exit 75
75 Categories of Processes Control in State and Capabilities Control E 1 E m Relation Observation Relation(E 1,, E m ) Relation Observation Processes Input: Data: Entities with properties Output: Data: List of relations Events: Detection or Loss of a relation 76
76 Categories of Processes Control in State and Capabilities E 1 E m Control Composition Observation Events Composite Objects Composition Observation Processes Input: Data: Entities with properties Output: Data: List of Composite Objects with CF Events: Detection or loss of a composite object 77
77 Example: Hand and Face Observer FaceAndHand Observer Control Control Events Video Entity Detection and Tracking E 1 E m Entity Composition Torso, Face, Hands Composed of Entity Tracker: Background difference and color Entity Grouper: Assigns roles to blobs 78
78 Example: Hand and Face Observer 79
79 Supervised Perceptual Process Events Configuration Requests for state Events Current State Response to commands Autonomic Supervisor Time ROI, S, Detection Method Detection Prediction ROI, S, Detection Method Video Stream ObservationMo ObservationMo dules Observation dules Modules Estimation Entities Intepretation Actors Observation Modules: Color Histogram Ratio Local Appearance Background Difference Motion History Image Local Appearance is described by Receptive Fields 96
80 What does it mean to see? Outline: The Science of Computer Vision Early Paradigms Geometric Foundations Active Vision Physics Based Vision: Color Models Real Time Tracking Systems Appearance Based Vision Possible Future Paradigms and Conclusions 97
81 Appearance The set of all possible images of an object, scene or event. 98
82 The Appearance Manifold The pixels of an image define a vector. The space of all images of an object is a manifold (the Appearance Manifold). The dimensions of the appearance manifold are the parameters of image formation problem: representing the appearance manifold 99
83 Plenoptic Dimensions Plenoptic function: A(x, y,, R, s, ) x, y - image position ϕ, γ - Lattitude and Longitude (2) R - Radius of view sphere (1) s - image scale factor (1) θ - Image plane rotation (1) Λ - Illumination (2 or more) 100
84 Sampled Plenoptic Dimensions Plenoptic image function: A(i, j, m, n,, s) i, j, Image Coordinates m, n - Lattitude and Longitude s - image scale factor (includes R, view sphere radius) θ - Image plane rotation Λ - Assume constant Illumination 101
85 Receptive Field Manifolds r M m,n (i,j) =< A(i,j,m,n), r ϑ (i,j;σ,θ) > ϑ k : M k : Vector or receptive fields Vector of local features for indexation and recognition. 102
86 Receptive Field Manifolds r M m,n (i,j) =< A(i,j,m,n), r ϑ (i,j;σ,θ) > Problem: Define i,j the receptive field functions 103
87 Chromatic Appearance L C 1 C 2 = R G B Transform color images (R, G, B) (L, C 1, C 2 ) 105
88 Chromatic receptive fields Puminance: information about object shape Chrominance: signature for recognition Luminance r - g r + g - b chrominance 106
89 Gaussian RF s are Steerable in Orientation Intrinsic orientation θ i (i,j) = Tan 1 ( < G y A(i,j) > < G x A(i,j) > ) Receptive field response at intrinsic orientation. < G x θ A(i, j) > = < G x A(i,j) > Cos(θ)+ < G y A(i,j) >Sin(θ) 107
90 Gaussian RF s are Steerable in Scale The intrinsic scale is an invariant for appearance σ (i, j) = Arg Max{< 2 G(σ) A(i,j) >} i σ 108
91 Probabilistic Recognition using Receptive Field Histograms Feature Vector: Probability of local appearance Probability of appearance given class C v k (x,y) =< ϕ k,a(x,y) > p( r v ) 1 N Tot h Tot ( r v ) p( r v C) 1 N C h C ( r v ) Probabilty of class C p(c) N C N Tot Probability of class C given appearance p(c r v ) = p(r v C)p(C) p( r v ) h (r C v ) h Tot ( r v ) 109
92 View Invariant Person Detection 110
93 View Invariant Person Detection 111
94 Local Appearance Manifold r V (i, j)=< A(i,j), r ϑ > A region of an image is a surface in receptive field space. The set of images of an object is a manifold Position in receptive field space allows pixel level matching 112
95 Recognition by prediction-verification Establish correspondance with most salient point. Propagate correspondance to neighboring pixels. 113
96 Local Appearance 114
97 View Invariant Person Recognition 115
98 What does it mean to see? Outline: The Science of Computer Vision Early Paradigms Geometric Foundations Active Vision Physics Based Vision: Color Models Real Time Tracking Systems Appearance Based Vision Possible Future Paradigms and Conclusions 116
99 Lessons from Biology Seeing is reaction to visual stimulus. Within verterbrates Visual Perception is organised as a series of associations mediated by filters Fixation is mediated in the superior colliculus Attention is mediated by a range of structures from the brainstem through the visual cortex to the frontal cortex. Vision develops as part of a sensor-moteur system whose primary purpose is homeostasis (internal and external). Biological vision is very specific to the structure of the organism and the nature of the environment. 117
100 Possible Future Paradigms Biologically Inspired Vision Insect Vision (Franceschini et al) Vertebrate Vision (Imitate human visual architecture) (cf Current IST-FET Call for Proposal) Ambient Perception (part of ambient intelligence) Adhoc network of large numbers of embedded devices with communication, sensing, display and computing 118
101 Conclusions: What does it mean to see? Human vision is the reaction to visual stimuli Visual skills formed by experience Visual reactions are mediated by experience and goals Human vision is part of sensori-motor and socialogical interaction. The engineering science of machine vision requires Foundations from geometry and signal analysis Techniques for learning visual skills Software engineering techniques for integration and control A theory of vision systems. The field is evolving rapidly, but we have far to go. 119
102 What does it mean to see? Outline: The Science of Computer Vision Early Paradigms Geometric Foundations Active Vision Physics Based Vision: Color Models Real Time Tracking Systems Appearance Based Vision Possible Future Paradigms and Conclusions 120
103 What does it mean to see? The engineering science of computer vision James L. Crowley Professor, I.N.P. Grenoble Projet PRIMA - Laboratory GRAVIR INRIA Rhône Alpes Grenoble, France 121
104 The Conferences and Journals of Computer Vision Journals: IJCV: International Journal of Computer Vision PAMI: IEEE Transactions on Pattern Analysis and Machine Intelligence IVC: Image and Vision Computing Conferences ICCV: International Conference on Computer Vision ECCV: European Conference on Computer Vision CVPR: IEEE (North American) Conferences on Computer Vision ICVS: International Conference on Vision Systems 122
Image Analysis and Formation (Formation et Analyse d'images)
Image Analysis and Formation (Formation et Analyse d'images) James L. Crowley ENSIMAG 3 - MMIS Option MIRV First Semester 2010/2011 Lesson 4 19 Oct 2010 Lesson Outline: 1 The Physics of Light...2 1.1 Photons
More informationComputer Vision James L. Crowley
Computer Vision James L. Crowley M2R MoSIG option GVR Lesson 1 Fall Semester 7 October 2008 Lesson Outline: Color Perception in Man and Machine 1 The Physics of Light...2 1.1 Photons and the Electo-Magnetic
More informationComputer Vision James L. Crowley
M2R MoSIG option GVR Lesson Outline: Computer Vision James L. Crowley Lesson 3 Visual Perception in Man and Machine Fall Semester 12 October 2012 1 The Physics of Light...2 1.1 Photons and the Electo-Magnetic
More informationRepresenting the World
Table of Contents Representing the World...1 Sensory Transducers...1 The Lateral Geniculate Nucleus (LGN)... 2 Areas V1 to V5 the Visual Cortex... 2 Computer Vision... 3 Intensity Images... 3 Image Focusing...
More informationThe SIFT (Scale Invariant Feature
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor developed by David Lowe University of British Columbia Initial paper ICCV 1999 Newer journal paper IJCV 2004 Review: Matt Brown s Canonical
More informationRecap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?
Recap: Features and filters Epipolar geometry & stereo vision Tuesday, Oct 21 Kristen Grauman UT-Austin Transforming and describing images; textures, colors, edges Recap: Grouping & fitting Now: Multiple
More informationDynamic Modeling of Free-Space for a Mobile Robot
Dynamic Modeling of Free-Space for a Mobile Robot 1. Introduction James L. Crowley, Patrick Reignier and Philippe Bobet LIFIA (IMAG) Institut National Polytechnique de Grenoble Grenoble, France Free-space
More informationEuclidean Reconstruction Independent on Camera Intrinsic Parameters
Euclidean Reconstruction Independent on Camera Intrinsic Parameters Ezio MALIS I.N.R.I.A. Sophia-Antipolis, FRANCE Adrien BARTOLI INRIA Rhone-Alpes, FRANCE Abstract bundle adjustment techniques for Euclidean
More informationCS4733 Class Notes, Computer Vision
CS4733 Class Notes, Computer Vision Sources for online computer vision tutorials and demos - http://www.dai.ed.ac.uk/hipr and Computer Vision resources online - http://www.dai.ed.ac.uk/cvonline Vision
More informationMETRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS
METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires
More informationMotion Tracking and Event Understanding in Video Sequences
Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!
More informationColor. making some recognition problems easy. is 400nm (blue) to 700 nm (red) more; ex. X-rays, infrared, radio waves. n Used heavily in human vision
Color n Used heavily in human vision n Color is a pixel property, making some recognition problems easy n Visible spectrum for humans is 400nm (blue) to 700 nm (red) n Machines can see much more; ex. X-rays,
More informationOmni Stereo Vision of Cooperative Mobile Robots
Omni Stereo Vision of Cooperative Mobile Robots Zhigang Zhu*, Jizhong Xiao** *Department of Computer Science **Department of Electrical Engineering The City College of the City University of New York (CUNY)
More informationLast update: May 4, Vision. CMSC 421: Chapter 24. CMSC 421: Chapter 24 1
Last update: May 4, 200 Vision CMSC 42: Chapter 24 CMSC 42: Chapter 24 Outline Perception generally Image formation Early vision 2D D Object recognition CMSC 42: Chapter 24 2 Perception generally Stimulus
More informationComputer Vision I - Appearance-based Matching and Projective Geometry
Computer Vision I - Appearance-based Matching and Projective Geometry Carsten Rother 05/11/2015 Computer Vision I: Image Formation Process Roadmap for next four lectures Computer Vision I: Image Formation
More informationThree-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras
Three-Dimensional Sensors Lecture 2: Projected-Light Depth Cameras Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inria.fr http://perception.inrialpes.fr/ Outline The geometry of active stereo.
More informationColor and Shading. Color. Shapiro and Stockman, Chapter 6. Color and Machine Vision. Color and Perception
Color and Shading Color Shapiro and Stockman, Chapter 6 Color is an important factor for for human perception for object and material identification, even time of day. Color perception depends upon both
More informationLocal qualitative shape from stereo. without detailed correspondence. Extended Abstract. Shimon Edelman. Internet:
Local qualitative shape from stereo without detailed correspondence Extended Abstract Shimon Edelman Center for Biological Information Processing MIT E25-201, Cambridge MA 02139 Internet: edelman@ai.mit.edu
More informationMiniature faking. In close-up photo, the depth of field is limited.
Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg
More informationFace detection in a video sequence - a temporal approach
Face detection in a video sequence - a temporal approach K. Mikolajczyk R. Choudhury C. Schmid INRIA Rhône-Alpes GRAVIR-CNRS, 655 av. de l Europe, 38330 Montbonnot, France {Krystian.Mikolajczyk,Ragini.Choudhury,Cordelia.Schmid}@inrialpes.fr
More informationMathematics of a Multiple Omni-Directional System
Mathematics of a Multiple Omni-Directional System A. Torii A. Sugimoto A. Imiya, School of Science and National Institute of Institute of Media and Technology, Informatics, Information Technology, Chiba
More informationProject 4 Results. Representation. Data. Learning. Zachary, Hung-I, Paul, Emanuel. SIFT and HoG are popular and successful.
Project 4 Results Representation SIFT and HoG are popular and successful. Data Hugely varying results from hard mining. Learning Non-linear classifier usually better. Zachary, Hung-I, Paul, Emanuel Project
More informationMarcel Worring Intelligent Sensory Information Systems
Marcel Worring worring@science.uva.nl Intelligent Sensory Information Systems University of Amsterdam Information and Communication Technology archives of documentaries, film, or training material, video
More informationDD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication
DD2423 Image Analysis and Computer Vision IMAGE FORMATION Mårten Björkman Computational Vision and Active Perception School of Computer Science and Communication November 8, 2013 1 Image formation Goal:
More informationRobot vision review. Martin Jagersand
Robot vision review Martin Jagersand What is Computer Vision? Computer Graphics Three Related fields Image Processing: Changes 2D images into other 2D images Computer Graphics: Takes 3D models, renders
More informationProc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp , Kobe, Japan, September 1992
Proc. Int. Symp. Robotics, Mechatronics and Manufacturing Systems 92 pp.957-962, Kobe, Japan, September 1992 Tracking a Moving Object by an Active Vision System: PANTHER-VZ Jun Miura, Hideharu Kawarabayashi,
More informationCapturing, Modeling, Rendering 3D Structures
Computer Vision Approach Capturing, Modeling, Rendering 3D Structures Calculate pixel correspondences and extract geometry Not robust Difficult to acquire illumination effects, e.g. specular highlights
More informationHow to Compute the Pose of an Object without a Direct View?
How to Compute the Pose of an Object without a Direct View? Peter Sturm and Thomas Bonfort INRIA Rhône-Alpes, 38330 Montbonnot St Martin, France {Peter.Sturm, Thomas.Bonfort}@inrialpes.fr Abstract. We
More informationProbabilistic Tracking and Reconstruction of 3D Human Motion in Monocular Video Sequences
Probabilistic Tracking and Reconstruction of 3D Human Motion in Monocular Video Sequences Presentation of the thesis work of: Hedvig Sidenbladh, KTH Thesis opponent: Prof. Bill Freeman, MIT Thesis supervisors
More informationA Simple Vision System
Chapter 1 A Simple Vision System 1.1 Introduction In 1966, Seymour Papert wrote a proposal for building a vision system as a summer project [4]. The abstract of the proposal starts stating a simple goal:
More informationAll human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them.
All human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them. - Aristotle University of Texas at Arlington Introduction
More informationComputer Vision Lecture 17
Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester
More informationDigital Image Processing COSC 6380/4393
Digital Image Processing COSC 6380/4393 Lecture 21 Nov 16 th, 2017 Pranav Mantini Ack: Shah. M Image Processing Geometric Transformation Point Operations Filtering (spatial, Frequency) Input Restoration/
More informationCOMPUTER VISION. Dr. Sukhendu Das Deptt. of Computer Science and Engg., IIT Madras, Chennai
COMPUTER VISION Dr. Sukhendu Das Deptt. of Computer Science and Engg., IIT Madras, Chennai 600036. Email: sdas@iitm.ac.in URL: //www.cs.iitm.ernet.in/~sdas 1 INTRODUCTION 2 Human Vision System (HVS) Vs.
More informationComputer Vision Lecture 17
Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week
More informationFeature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies
Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies M. Lourakis, S. Tzurbakis, A. Argyros, S. Orphanoudakis Computer Vision and Robotics Lab (CVRL) Institute of
More informationSensor Modalities. Sensor modality: Different modalities:
Sensor Modalities Sensor modality: Sensors which measure same form of energy and process it in similar ways Modality refers to the raw input used by the sensors Different modalities: Sound Pressure Temperature
More informationSimultaneous Vanishing Point Detection and Camera Calibration from Single Images
Simultaneous Vanishing Point Detection and Camera Calibration from Single Images Bo Li, Kun Peng, Xianghua Ying, and Hongbin Zha The Key Lab of Machine Perception (Ministry of Education), Peking University,
More informationVisual Recognition: Image Formation
Visual Recognition: Image Formation Raquel Urtasun TTI Chicago Jan 5, 2012 Raquel Urtasun (TTI-C) Visual Recognition Jan 5, 2012 1 / 61 Today s lecture... Fundamentals of image formation You should know
More informationLocal features: detection and description. Local invariant features
Local features: detection and description Local invariant features Detection of interest points Harris corner detection Scale invariant blob detection: LoG Description of local patches SIFT : Histograms
More informationPhysics-based Vision: an Introduction
Physics-based Vision: an Introduction Robby Tan ANU/NICTA (Vision Science, Technology and Applications) PhD from The University of Tokyo, 2004 1 What is Physics-based? An approach that is principally concerned
More informationFeature descriptors and matching
Feature descriptors and matching Detections at multiple scales Invariance of MOPS Intensity Scale Rotation Color and Lighting Out-of-plane rotation Out-of-plane rotation Better representation than color:
More informationEECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline
EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT Oct. 15, 2013 Prof. Ronald Fearing Electrical Engineering and Computer Sciences University of California, Berkeley (slides courtesy of Prof. John Wawrzynek)
More informationMulti-Modal Tracking of Faces for Video Communications
Multi-Modal Tracking of Faces for Video Communications James L. Crowley and Francois Berard GRAVIR - IMAG, I.N.P. Grenoble 46 Ave Félix Viallet 38031 Grenoble, France Abstract This paper describes a system
More informationFlexible Calibration of a Portable Structured Light System through Surface Plane
Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured
More informationBinocular Stereo Vision. System 6 Introduction Is there a Wedge in this 3D scene?
System 6 Introduction Is there a Wedge in this 3D scene? Binocular Stereo Vision Data a stereo pair of images! Given two 2D images of an object, how can we reconstruct 3D awareness of it? AV: 3D recognition
More informationBrand Identification Using Gaussian Derivative Histograms
Machine Vision and Applications manuscript No. (will be inserted by the editor) Brand Identification Using Gaussian Derivative Histograms Daniela Hall, Fabien Pélisson, Olivier Riff, James L. Crowley PRIMA
More informationComputer Vision I - Appearance-based Matching and Projective Geometry
Computer Vision I - Appearance-based Matching and Projective Geometry Carsten Rother 01/11/2016 Computer Vision I: Image Formation Process Roadmap for next four lectures Computer Vision I: Image Formation
More informationAn Overview of Matchmoving using Structure from Motion Methods
An Overview of Matchmoving using Structure from Motion Methods Kamyar Haji Allahverdi Pour Department of Computer Engineering Sharif University of Technology Tehran, Iran Email: allahverdi@ce.sharif.edu
More informationCOSC579: Scene Geometry. Jeremy Bolton, PhD Assistant Teaching Professor
COSC579: Scene Geometry Jeremy Bolton, PhD Assistant Teaching Professor Overview Linear Algebra Review Homogeneous vs non-homogeneous representations Projections and Transformations Scene Geometry The
More informationStereo Vision. MAN-522 Computer Vision
Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in
More informationHomographies and RANSAC
Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position
More informationCAP 5415 Computer Vision Fall 2012
CAP 5415 Computer Vision Fall 01 Dr. Mubarak Shah Univ. of Central Florida Office 47-F HEC Lecture-5 SIFT: David Lowe, UBC SIFT - Key Point Extraction Stands for scale invariant feature transform Patented
More informationIntrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting
Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting R. Maier 1,2, K. Kim 1, D. Cremers 2, J. Kautz 1, M. Nießner 2,3 Fusion Ours 1
More informationLaser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR
Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and
More informationDepth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy
Depth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy Sharjeel Anwar, Dr. Shoaib, Taosif Iqbal, Mohammad Saqib Mansoor, Zubair
More informationComputational Foundations of Cognitive Science
Computational Foundations of Cognitive Science Lecture 16: Models of Object Recognition Frank Keller School of Informatics University of Edinburgh keller@inf.ed.ac.uk February 23, 2010 Frank Keller Computational
More informationCorrespondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri] Introduction Disparity: Informally: difference between two pictures Allows us to gain a strong
More informationLocal invariant features
Local invariant features Tuesday, Oct 28 Kristen Grauman UT-Austin Today Some more Pset 2 results Pset 2 returned, pick up solutions Pset 3 is posted, due 11/11 Local invariant features Detection of interest
More informationURBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES
URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES An Undergraduate Research Scholars Thesis by RUI LIU Submitted to Honors and Undergraduate Research Texas A&M University in partial fulfillment
More informationImage Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania
Image Formation Antonino Furnari Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania furnari@dmi.unict.it 18/03/2014 Outline Introduction; Geometric Primitives
More information9.913 Pattern Recognition for Vision. Class I - Overview. Instructors: B. Heisele, Y. Ivanov, T. Poggio
9.913 Class I - Overview Instructors: B. Heisele, Y. Ivanov, T. Poggio TOC Administrivia Problems of Computer Vision and Pattern Recognition Overview of classes Quick review of Matlab Administrivia Instructors:
More informationImage Formation. Ed Angel Professor of Computer Science, Electrical and Computer Engineering, and Media Arts University of New Mexico
Image Formation Ed Angel Professor of Computer Science, Electrical and Computer Engineering, and Media Arts University of New Mexico 1 Objectives Fundamental imaging notions Physical basis for image formation
More information(0, 1, 1) (0, 1, 1) (0, 1, 0) What is light? What is color? Terminology
lecture 23 (0, 1, 1) (0, 0, 0) (0, 0, 1) (0, 1, 1) (1, 1, 1) (1, 1, 0) (0, 1, 0) hue - which ''? saturation - how pure? luminance (value) - intensity What is light? What is? Light consists of electromagnetic
More informationTracking of Human Body using Multiple Predictors
Tracking of Human Body using Multiple Predictors Rui M Jesus 1, Arnaldo J Abrantes 1, and Jorge S Marques 2 1 Instituto Superior de Engenharia de Lisboa, Postfach 351-218317001, Rua Conselheiro Emído Navarro,
More informationAugmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit
Augmented Reality VU Computer Vision 3D Registration (2) Prof. Vincent Lepetit Feature Point-Based 3D Tracking Feature Points for 3D Tracking Much less ambiguous than edges; Point-to-point reprojection
More informationIntroduction to Computer Graphics with WebGL
Introduction to Computer Graphics with WebGL Ed Angel Professor Emeritus of Computer Science Founding Director, Arts, Research, Technology and Science Laboratory University of New Mexico Image Formation
More informationA Summary of Projective Geometry
A Summary of Projective Geometry Copyright 22 Acuity Technologies Inc. In the last years a unified approach to creating D models from multiple images has been developed by Beardsley[],Hartley[4,5,9],Torr[,6]
More informationLecture 1 Image Formation.
Lecture 1 Image Formation peimt@bit.edu.cn 1 Part 3 Color 2 Color v The light coming out of sources or reflected from surfaces has more or less energy at different wavelengths v The visual system responds
More informationMotion Estimation and Optical Flow Tracking
Image Matching Image Retrieval Object Recognition Motion Estimation and Optical Flow Tracking Example: Mosiacing (Panorama) M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 Example 3D Reconstruction
More informationNoise Model. Important Noise Probability Density Functions (Cont.) Important Noise Probability Density Functions
Others -- Noise Removal Techniques -- Edge Detection Techniques -- Geometric Operations -- Color Image Processing -- Color Spaces Xiaojun Qi Noise Model The principal sources of noise in digital images
More informationPop Quiz 1 [10 mins]
Pop Quiz 1 [10 mins] 1. An audio signal makes 250 cycles in its span (or has a frequency of 250Hz). How many samples do you need, at a minimum, to sample it correctly? [1] 2. If the number of bits is reduced,
More informationAn introduction to 3D image reconstruction and understanding concepts and ideas
Introduction to 3D image reconstruction An introduction to 3D image reconstruction and understanding concepts and ideas Samuele Carli Martin Hellmich 5 febbraio 2013 1 icsc2013 Carli S. Hellmich M. (CERN)
More informationCatadioptric camera model with conic mirror
LÓPEZ-NICOLÁS, SAGÜÉS: CATADIOPTRIC CAMERA MODEL WITH CONIC MIRROR Catadioptric camera model with conic mirror G. López-Nicolás gonlopez@unizar.es C. Sagüés csagues@unizar.es Instituto de Investigación
More informationMulti-view stereo. Many slides adapted from S. Seitz
Multi-view stereo Many slides adapted from S. Seitz Beyond two-view stereo The third eye can be used for verification Multiple-baseline stereo Pick a reference image, and slide the corresponding window
More informationCOMPUTER AND ROBOT VISION
VOLUME COMPUTER AND ROBOT VISION Robert M. Haralick University of Washington Linda G. Shapiro University of Washington T V ADDISON-WESLEY PUBLISHING COMPANY Reading, Massachusetts Menlo Park, California
More informationFace Recognition At-a-Distance Based on Sparse-Stereo Reconstruction
Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,
More informationSelf-calibration of a pair of stereo cameras in general position
Self-calibration of a pair of stereo cameras in general position Raúl Rojas Institut für Informatik Freie Universität Berlin Takustr. 9, 14195 Berlin, Germany Abstract. This paper shows that it is possible
More informationStereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman
Stereo 11/02/2012 CS129, Brown James Hays Slides by Kristen Grauman Multiple views Multi-view geometry, matching, invariant features, stereo vision Lowe Hartley and Zisserman Why multiple views? Structure
More informationThink-Pair-Share. What visual or physiological cues help us to perceive 3D shape and depth?
Think-Pair-Share What visual or physiological cues help us to perceive 3D shape and depth? [Figure from Prados & Faugeras 2006] Shading Focus/defocus Images from same point of view, different camera parameters
More informationPerception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich.
Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Perception Sensors Vision Uncertainties, Line extraction
More information3D graphics, raster and colors CS312 Fall 2010
Computer Graphics 3D graphics, raster and colors CS312 Fall 2010 Shift in CG Application Markets 1989-2000 2000 1989 3D Graphics Object description 3D graphics model Visualization 2D projection that simulates
More informationIllumination and Shading
Illumination and Shading Light sources emit intensity: assigns intensity to each wavelength of light Humans perceive as a colour - navy blue, light green, etc. Exeriments show that there are distinct I
More informationEpipolar geometry contd.
Epipolar geometry contd. Estimating F 8-point algorithm The fundamental matrix F is defined by x' T Fx = 0 for any pair of matches x and x in two images. Let x=(u,v,1) T and x =(u,v,1) T, each match gives
More informationBehavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism
Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Sho ji Suzuki, Tatsunori Kato, Minoru Asada, and Koh Hosoda Dept. of Adaptive Machine Systems, Graduate
More informationUnderstanding Variability
Understanding Variability Why so different? Light and Optics Pinhole camera model Perspective projection Thin lens model Fundamental equation Distortion: spherical & chromatic aberration, radial distortion
More informationAnno accademico 2006/2007. Davide Migliore
Robotica Anno accademico 6/7 Davide Migliore migliore@elet.polimi.it Today What is a feature? Some useful information The world of features: Detectors Edges detection Corners/Points detection Descriptors?!?!?
More informationSegmentation and Tracking of Partial Planar Templates
Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract
More informationRobotics Programming Laboratory
Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car
More informationTracking and Recognizing People in Colour using the Earth Mover s Distance
Tracking and Recognizing People in Colour using the Earth Mover s Distance DANIEL WOJTASZEK, ROBERT LAGANIÈRE S.I.T.E. University of Ottawa, Ottawa, Ontario, Canada K1N 6N5 danielw@site.uottawa.ca, laganier@site.uottawa.ca
More informationAnnouncements. Hough Transform [ Patented 1962 ] Generalized Hough Transform, line fitting. Assignment 2: Due today Midterm: Thursday, May 5 in class
Announcements Generalized Hough Transform, line fitting Assignment 2: Due today Midterm: Thursday, May 5 in class Introduction to Computer Vision CSE 152 Lecture 11a What is region like if: 1. λ 1 = 0?
More informationCSE 4392/5369. Dr. Gian Luca Mariottini, Ph.D.
University of Texas at Arlington CSE 4392/5369 Introduction to Vision Sensing Dr. Gian Luca Mariottini, Ph.D. Department of Computer Science and Engineering University of Texas at Arlington WEB : http://ranger.uta.edu/~gianluca
More informationDept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan
An Application of Vision-Based Learning for a Real Robot in RoboCup - A Goal Keeping Behavior for a Robot with an Omnidirectional Vision and an Embedded Servoing - Sho ji Suzuki 1, Tatsunori Kato 1, Hiroshi
More informationMultimedia Computing: Algorithms, Systems, and Applications: Edge Detection
Multimedia Computing: Algorithms, Systems, and Applications: Edge Detection By Dr. Yu Cao Department of Computer Science The University of Massachusetts Lowell Lowell, MA 01854, USA Part of the slides
More informationSUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS
SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS Cognitive Robotics Original: David G. Lowe, 004 Summary: Coen van Leeuwen, s1460919 Abstract: This article presents a method to extract
More informationTarget Tracking Using Mean-Shift And Affine Structure
Target Tracking Using Mean-Shift And Affine Structure Chuan Zhao, Andrew Knight and Ian Reid Department of Engineering Science, University of Oxford, Oxford, UK {zhao, ian}@robots.ox.ac.uk Abstract Inthispaper,wepresentanewapproachfortracking
More informationComputer Vision Course Lecture 02. Image Formation Light and Color. Ceyhun Burak Akgül, PhD cba-research.com. Spring 2015 Last updated 04/03/2015
Computer Vision Course Lecture 02 Image Formation Light and Color Ceyhun Burak Akgül, PhD cba-research.com Spring 2015 Last updated 04/03/2015 Photo credit: Olivier Teboul vision.mas.ecp.fr/personnel/teboul
More information3D Model Acquisition by Tracking 2D Wireframes
3D Model Acquisition by Tracking 2D Wireframes M. Brown, T. Drummond and R. Cipolla {96mab twd20 cipolla}@eng.cam.ac.uk Department of Engineering University of Cambridge Cambridge CB2 1PZ, UK Abstract
More informationCIS 580, Machine Perception, Spring 2015 Homework 1 Due: :59AM
CIS 580, Machine Perception, Spring 2015 Homework 1 Due: 2015.02.09. 11:59AM Instructions. Submit your answers in PDF form to Canvas. This is an individual assignment. 1 Camera Model, Focal Length and
More informationCHAPTER 9. Classification Scheme Using Modified Photometric. Stereo and 2D Spectra Comparison
CHAPTER 9 Classification Scheme Using Modified Photometric Stereo and 2D Spectra Comparison 9.1. Introduction In Chapter 8, even we combine more feature spaces and more feature generators, we note that
More information