EXAM SOLUTIONS. Computer Vision Course 2D1420 Thursday, 11 th of march 2003,
|
|
- Paula Owens
- 5 years ago
- Views:
Transcription
1 Numerical Analysis and Computer Science, KTH Danica Kragic EXAM SOLUTIONS Computer Vision Course 2D1420 Thursday, 11 th of march 2003, Exercise 1 (5*2=10 credits) Answer at most 5 of the following questions. If you answer more than five, only first five answers will be considered and the rest ignored. (1) What is fovea and what is its function? Describe what are cones and rods. Fovea: shallow pit close to blind spot with high concentration of photoreceptors, fraction of mm in diameter, supports high-resolution vision in bright light, controls gaze. Cones (tappar) : used for photopic (light) vision, concentrated in the fovea centralis of the macula, three color types - red, green, blue, require bright light, relatively sparse outside fovea, wavelength specific. Rods (stavar): used for scotopic or dark vision, distributed over the entire retina, with a peak at about 20deg from the fovea centralis, similar sensitivity wrt color (not wavelength specific), very sensitive to light (quantum level). Each eye contains about 120 million rods and 5 million cones. (2) Explain terms radiance and irradiance. Radiance: Amount of light radiated from a surface into a given solid angle per unit area. The area is the foreshortened area, as seen from the direction that the light is being emitted. Irradiance: Light power per unit area (watts per square meter) incident on a surface. If surface tilts away from light, same amount of light strikes bigger surface - foreshortening (less irradiance). (3) Explain terms sampling and quantization. Sampling: selection of a discrete grid to represent an image; Spatial deiscretization of an image Quantization: Mapping of the brightness into a numerical value; Assigning a physical measurement to one of a discrete set of points in a range. 1
2 (4) What are the common image points distance measures? Give at least two examples. Common distance measures: Euclidean distance City block distance Chessboard distance d(p,q) = (x u) 2 + (y v) 2 d(p,q) = x u + y v d(p,q) = max( x u, y v ) (5) What is ment by contrast reversal in terms of grey-level transformations? Draw the corresponding linear transformation. Contrast reversal - basically inverting the grey-level values. f (x) = g(1 x) (6) How is a signal in a Fourier domain affected by a) translation and b) mirroring of the original signal in the spectral domain. Translation affects only phase while power spectrum is translation invariant. Mirroring in spectral and spatial domain are the same. (7) What is ment by ideal low-pass filter? Is this filter suitable to use in terms of image processing? If yes, give an example of its application. If no, explain why. The ideal low-pass filter is a mathematically idealized version of a smoothing filter. In a frequencydomain representation of the image, all frequencies above a threshold F are discarded, i.e. this 2
3 filter passes low frequencies so image becomes blurred. This method of smoothing tends to create images with ringing at sharp boundaries in the picture which is its drawback. The observation that the application of the low-pass filter is equivalent to the convolution of the image with the sinc function provides an explanation for this phenomenon. This filter cannot be implemented in hardware. Ideal low pass filter in 2D: ĥ(v) = rect( v ) where v = v 2 1 2v + v2 2 c v c cut-off frequency Impulse response h(x) = 2πv 2 J 1 (2πv c x ) c 2πv c x J 1 - first order Bessel function (8) In what cases is spectral filtering more appropriate than spatial one? Give examples. 1) If the noise is periodic, 2) If we want to filter the image using large kernels. (9) What is ment by region growing? When is it commonly used? Region growing - homogeneous regions grows in size by including similar neighboring pixels, the final result does not necessarily need to cover the entire image First, seed regions have to be extracted, and these seed regions are iteratively grown at their borders by accepting new pixels being consistent with the pixels already being contained in the region. After each iteration, the homogeneity value of the region has to be re-calculated using also the new pixels. The results of region growing heavily depend of a proper selection of seed points Usage: divide image pixels into a set of classes based on local properties and features (classification), matching in stereo, split-and-merge transform etc. (10) What is region adjacency graph? What corresponds merging in terms of a region adjacency graph? It is used to store region attributes for every region - related to the level of homogeneity of regions. (Image provided in lectures.) merge = graph contraction 3
4 Exercise 2 ( =7 credits) (a) Describe basics and draw figures of perspective and orthographic camera models. Given a set of parallel lines in 3D - explain what is the difference in their image projections for both models. Perspective projection Orthographic projection x f = X Z, x = X, y f = Y Z y = Y For the orthographic camera model, parallel lines remain parallel and for a perspective camera model the lines intersect in a vanishing point. Image Image focal point Perspective projection Orthographic projection (b) Under what conditions will a set of parallel lines viewed with a pinhole camera have its vanishing point (in the image) at infinity? In case when lines are in the plane parallel to the image plane. (c) If the area of a planar square in the scene is A, what is it s area in an image under orthographic projection? Give your answer in terms of any parameters neccessary to define this relationsship, specifying what each parameter means. A cos θ where θ is the angle between the surface normal and the optical axis. 4
5 (d) What is the difference between an affine and perspective camera model? Affine camera can be seen as a linear model, while perspective camera is a non-linear one. Affine preserves parallel lines, and perspective does not. Affine camera is a good model when all the points lie on a relatively planar surface and their relative depth from the camera is large. Afiine camera - 8 prameters, perspective camera 12 parameters. Exercise 3 ( =9 credits) (a) An image has been smoothed using following kernel: k [ ]. Can repeated convolutions of an image with the kernel 1 [1 1] 2 be used to otain the same result as with the first kernel. If yes, how many convolutions are needed? If no, explain the reasons why. What should be the constant k so that the filter gain is equal to 1? Yes, after 4 convolutions we obtain the same filter and k = 1 32 since =32. (b) Show that the kernel k [ ] can be written as multiple convolutions of a low pass filter 1 2 [1 1] and a high pass filter [1 1]. What should k be in order for the kernel to be a good approximation of the first order derivative? What is the frequency response of this kernel? Yes, since: [1 1] [1 1] = [1 2 1] [1 2 1] [1 1] = [ ] [ ] [1 1] = [ ] It is obvious that k = 1 8 because 3 convolutions with a low-pass filter have been performed and the maximum value should not change. Also, to obtain the kernel [ ] the sign has to be negative. 5
6 Frequency response (according to lectures): e i2ω 2e iω + 2e iω + e i2ω = 4 i sinω + 2 i sin2ω = i (4sinω + 2 sin2ω) (c) Give an example of a mean (unweighted averaging) filter. Is a mean filter in general separable? Why do we prefer separable filters? Let us take the simplest example: Since unweighted averaging is assumed, all elements in the matrix are equal. It is always separable since its rank is equal to one. Separable filters are prefered since they can be implemented as repeated convolutions with one-dimensional kernels. (d) Give an image below before (left) and after (right) a smoothing filter was applied. The size of the filter is shown as a small square in the upper-left corner in the image (as you can see, its size is rather small compared to the image size). In your opinion, which one of the following filter types most likely produced the image on the right: 1) mean (unweighted averaging) filter, 2) median filter, or 3) Gaussian filter. Motivate your answer.. Median filter since the line on the left of the image dissapperas - with any other filter it would remain and be extended. 6
7 Exercise 4 (2+2+1=5 credits) You are given an image of an object on a background containing a strong illumination gradient. You are supposed to segment the object from the background in order to estimate its moment descriptors. (a) Sketch the histogram of the image and explain what are the problems related to choosing a suitable threshold. The histogram is not clearly bimodal - difficult to obtain local minimum (b) Propose a suitable methodology that could be used to perform successful segmentation. a) Local adaptive thresholding selects an individual threshold for each pixel based on the range of intensity values in its local neighbourhood. This allows for thresholding of an image whose global intensity histogram doesn t contain distinctive peaks. b) estimating the illumination gradient and 7
8 substracting it from the image c) estimating gradient close to the object boundaries and using this to perform adaptive thresholding (c) You have applied histogram equalization to an image. If you apply histogram equalization to an image second time, will you improve image quality even more? Motivate your aswer. No since the first histogram equalization operation will flatten the image as much as possible. The second application will not change the histogram significantly. Exercise 5 (2+2+2=6 credits) (a) What is the epipolar constraint and how can it be used in stereo matching? Represents geometry of two cameras, reduces a correspondance problem to 1D search along an epipolar line. A point in one view generates an epipolar line in the other view. The corresponding point lies on this line. Epipolar geometry is a result of coplanarity between camera centers and a world point - all of them lie in an epipolar plane. X q q o p p o centers of projection (b) Assume a pair of parallel cameras with their optical axes perpendicular to the baseline. How do the epipolar lines look like? Where are the epipoles for this type of camera system? The epipolar lines are parallel to each other (also, in the direction of th image y axis). If the optical centers of the camera are on the same height, the corresponding epipolar lines will be on the same height - one line for both images. The epipoles are in the infinity. (c) Estimate the essential matrix between two consecutive images for a forward translating camera. What is the equation of the epipolar line for the point p=[x y 1]? In general E = t S R where t S is a skew-symmetric matrix related to translation vector. For a forward translating camera (see figure), we have 8
9 R = I, and t S = 0 t Z 0 t Z therefore E = 0 t Z 0 t Z From l = E p the epipolar line for a point p = [x y 1] is l = 0 t Z 0 t Z x y 1 = t Z y t Z x 0 9
10 Exercise 6 (2+2+3=7 credits) (a) What is the difference between motion field and optical flow? Motion field: Projection of 3D motion onto the image plane. Optical flow: Apparent motion of the image brightness pattern (not necessarily related to the motion field). (b) Under what assumptions does the optical flow constraint work? When does it not work? Brightness constancy: does not work if the light source moves, while the object is static. Textured region: does not working without sufficient local image structure. alt. Lambertian assumption: might not work if objects have specularities (mirror-like objects). alt. Spatially local constancy: necessary in practice, since derivatives have to be determined within some region. (c) Assume that an object located at a distance 10 m is moving at a speed of 3 m/s in a direction parallel to the image plane as indicated in the figure. How fast should the camera be rotated around the x and y-axis, such that the object remains fixated in the centre of the image. Assume that the focal length is f = 5. y 30 o x The motion of the object in 3D is { Ẋ = 3cos30 ms 1 = 3 3/2 ms ms 1 Ẏ = 3sin30 ms 1 = 3/2 ms 1 = 1.5 ms 1 The translational motion field is given by { ut = f Ẋ/Z = f 3 3/20 s 1 v t = f Ẏ /Z = f 3/20 s 1 and the rotational motion field is { ur = f ω y v r = f ω x Since { motion field is supposed { to be zero u = ut + u r = 0 v = v t + v r = 0 ur = u t = f ω y v r = v t = f ω x 10
11 the rotation of the camera has to be { ωx = v t / f = 3/20 s 1 = 0.15 s s 1 ω y = u t / f = 3 3/20 s s s 1 Exercise 7 (3+3=6 credits) Assume that we want to classify an image into one of two classes: C A and C B. We know that the prior probability for A is two time as large as B: p(a) = 2 3 and p(b) = 1 3. After the preprocessing step we get a feature map z. Calculate an expression for a Bayesian classifier: (a) in the case of a one dimensional feature map, σ A = 4, σ B = 1, and m A = m B = 0, and: p(z C k ) = 1 e (z m k) 2 /(2σ 2 k ). 2πσk Sketch the decision functions and calculate the decision boundaries Using Bayes rule (withouth the normalization), we have: p(c A )p(z C A ) = p(c B )p(z C B ) Using the above equation for the PDFs and priors: 2 3 2π 4 e z2 /32 = 1 3 2π e z2 /2 From which we get: Decision boundaries given by: Decision function given by: z = ±1.216 z > z < Class A if z > Class B if z <
12 (b) in the( case of a) two dimensional ( ) feature map, Σ A =, Σ 0 2 B =, m 0 1 A = m B = 0, and p(z C k ) = Calculate the two dimensional decision boundary. From the above, we have: 1 2π detσ k 1/2 e (z m k) T Σ 1 k (z m k )/2. detσ A = 2, Σ 1 A = 1 ( detσ B = 2, Σ 1 B = 1 ( Again, using Bayes rule withouth normalization: ) ) 1 3π 2 e z T z 4 = 1 6π 2 e z T z 2 e (2x2 +y 2 )/4 = e (x2 +2y 2 )/4 ln2 (2x 2 + y 2 )/4 + (x 2 + 2y 2 )/4 = 0 Two-dimensional decision boundary reprsented by a hyperbola: x 2 y 2 = 4ln2 12
EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,
School of Computer Science and Communication, KTH Danica Kragic EXAM SOLUTIONS Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, 14.00 19.00 Grade table 0-25 U 26-35 3 36-45
More informationDD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication
DD2423 Image Analysis and Computer Vision IMAGE FORMATION Mårten Björkman Computational Vision and Active Perception School of Computer Science and Communication November 8, 2013 1 Image formation Goal:
More informationELEC Dr Reji Mathew Electrical Engineering UNSW
ELEC 4622 Dr Reji Mathew Electrical Engineering UNSW Review of Motion Modelling and Estimation Introduction to Motion Modelling & Estimation Forward Motion Backward Motion Block Motion Estimation Motion
More information1 (5 max) 2 (10 max) 3 (20 max) 4 (30 max) 5 (10 max) 6 (15 extra max) total (75 max + 15 extra)
Mierm Exam CS223b Stanford CS223b Computer Vision, Winter 2004 Feb. 18, 2004 Full Name: Email: This exam has 7 pages. Make sure your exam is not missing any sheets, and write your name on every page. The
More informationFinal Exam Study Guide
Final Exam Study Guide Exam Window: 28th April, 12:00am EST to 30th April, 11:59pm EST Description As indicated in class the goal of the exam is to encourage you to review the material from the course.
More informationChapter 3 Image Registration. Chapter 3 Image Registration
Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation
More informationFundamentals of Digital Image Processing
\L\.6 Gw.i Fundamentals of Digital Image Processing A Practical Approach with Examples in Matlab Chris Solomon School of Physical Sciences, University of Kent, Canterbury, UK Toby Breckon School of Engineering,
More informationUnderstanding Variability
Understanding Variability Why so different? Light and Optics Pinhole camera model Perspective projection Thin lens model Fundamental equation Distortion: spherical & chromatic aberration, radial distortion
More informationMultiple View Geometry
Multiple View Geometry CS 6320, Spring 2013 Guest Lecture Marcel Prastawa adapted from Pollefeys, Shah, and Zisserman Single view computer vision Projective actions of cameras Camera callibration Photometric
More informationComputer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier
Computer Vision 2 SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung Computer Vision 2 Dr. Benjamin Guthier 1. IMAGE PROCESSING Computer Vision 2 Dr. Benjamin Guthier Content of this Chapter Non-linear
More informationRange Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation
Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical
More informationCS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka
CS223b Midterm Exam, Computer Vision Monday February 25th, Winter 2008, Prof. Jana Kosecka Your name email This exam is 8 pages long including cover page. Make sure your exam is not missing any pages.
More informationCOSC579: Scene Geometry. Jeremy Bolton, PhD Assistant Teaching Professor
COSC579: Scene Geometry Jeremy Bolton, PhD Assistant Teaching Professor Overview Linear Algebra Review Homogeneous vs non-homogeneous representations Projections and Transformations Scene Geometry The
More informationIntroduction to Computer Vision. Introduction CMPSCI 591A/691A CMPSCI 570/670. Image Formation
Introduction CMPSCI 591A/691A CMPSCI 570/670 Image Formation Lecture Outline Light and Optics Pinhole camera model Perspective projection Thin lens model Fundamental equation Distortion: spherical & chromatic
More informationComputer Vision I - Basics of Image Processing Part 1
Computer Vision I - Basics of Image Processing Part 1 Carsten Rother 28/10/2014 Computer Vision I: Basics of Image Processing Link to lectures Computer Vision I: Basics of Image Processing 28/10/2014 2
More informationLecture 6: Edge Detection
#1 Lecture 6: Edge Detection Saad J Bedros sbedros@umn.edu Review From Last Lecture Options for Image Representation Introduced the concept of different representation or transformation Fourier Transform
More informationCapturing, Modeling, Rendering 3D Structures
Computer Vision Approach Capturing, Modeling, Rendering 3D Structures Calculate pixel correspondences and extract geometry Not robust Difficult to acquire illumination effects, e.g. specular highlights
More informationMidterm Exam Solutions
Midterm Exam Solutions Computer Vision (J. Košecká) October 27, 2009 HONOR SYSTEM: This examination is strictly individual. You are not allowed to talk, discuss, exchange solutions, etc., with other fellow
More informationFundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision
Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision What Happened Last Time? Human 3D perception (3D cinema) Computational stereo Intuitive explanation of what is meant by disparity Stereo matching
More informationStereo Vision. MAN-522 Computer Vision
Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in
More informationCS201 Computer Vision Lect 4 - Image Formation
CS201 Computer Vision Lect 4 - Image Formation John Magee 9 September, 2014 Slides courtesy of Diane H. Theriault Question of the Day: Why is Computer Vision hard? Something to think about from our view
More informationLaser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR
Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and
More informationFinal Review CMSC 733 Fall 2014
Final Review CMSC 733 Fall 2014 We have covered a lot of material in this course. One way to organize this material is around a set of key equations and algorithms. You should be familiar with all of these,
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationCorrespondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri] Introduction Disparity: Informally: difference between two pictures Allows us to gain a strong
More informationMotion Analysis. Motion analysis. Now we will talk about. Differential Motion Analysis. Motion analysis. Difference Pictures
Now we will talk about Motion Analysis Motion analysis Motion analysis is dealing with three main groups of motionrelated problems: Motion detection Moving object detection and location. Derivation of
More informationDigital Image Processing COSC 6380/4393
Digital Image Processing COSC 6380/4393 Lecture 4 Jan. 24 th, 2019 Slides from Dr. Shishir K Shah and Frank (Qingzhong) Liu Digital Image Processing COSC 6380/4393 TA - Office: PGH 231 (Update) Shikha
More informationStereo imaging ideal geometry
Stereo imaging ideal geometry (X,Y,Z) Z f (x L,y L ) f (x R,y R ) Optical axes are parallel Optical axes separated by baseline, b. Line connecting lens centers is perpendicular to the optical axis, and
More informationAll good things must...
Lecture 17 Final Review All good things must... UW CSE vision faculty Course Grading Programming Projects (80%) Image scissors (20%) -DONE! Panoramas (20%) - DONE! Content-based image retrieval (20%) -
More informationIntroduction to Computer Vision. Human Eye Sampling
Human Eye Sampling Sampling Rough Idea: Ideal Case 23 "Digitized Image" "Continuous Image" Dirac Delta Function 2D "Comb" δ(x,y) = 0 for x = 0, y= 0 s δ(x,y) dx dy = 1 f(x,y)δ(x-a,y-b) dx dy = f(a,b) δ(x-ns,y-ns)
More informationRepresenting the World
Table of Contents Representing the World...1 Sensory Transducers...1 The Lateral Geniculate Nucleus (LGN)... 2 Areas V1 to V5 the Visual Cortex... 2 Computer Vision... 3 Intensity Images... 3 Image Focusing...
More information2: Image Display and Digital Images. EE547 Computer Vision: Lecture Slides. 2: Digital Images. 1. Introduction: EE547 Computer Vision
EE547 Computer Vision: Lecture Slides Anthony P. Reeves November 24, 1998 Lecture 2: Image Display and Digital Images 2: Image Display and Digital Images Image Display: - True Color, Grey, Pseudo Color,
More informationVisual Recognition: Image Formation
Visual Recognition: Image Formation Raquel Urtasun TTI Chicago Jan 5, 2012 Raquel Urtasun (TTI-C) Visual Recognition Jan 5, 2012 1 / 61 Today s lecture... Fundamentals of image formation You should know
More informationHomogeneous Coordinates. Lecture18: Camera Models. Representation of Line and Point in 2D. Cross Product. Overall scaling is NOT important.
Homogeneous Coordinates Overall scaling is NOT important. CSED44:Introduction to Computer Vision (207F) Lecture8: Camera Models Bohyung Han CSE, POSTECH bhhan@postech.ac.kr (",, ) ()", ), )) ) 0 It is
More informationCamera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration
Camera Calibration Jesus J Caban Schedule! Today:! Camera calibration! Wednesday:! Lecture: Motion & Optical Flow! Monday:! Lecture: Medical Imaging! Final presentations:! Nov 29 th : W. Griffin! Dec 1
More informationMidterm Examination CS 534: Computational Photography
Midterm Examination CS 534: Computational Photography November 3, 2016 NAME: Problem Score Max Score 1 6 2 8 3 9 4 12 5 4 6 13 7 7 8 6 9 9 10 6 11 14 12 6 Total 100 1 of 8 1. [6] (a) [3] What camera setting(s)
More informationComparison between Motion Analysis and Stereo
MOTION ESTIMATION The slides are from several sources through James Hays (Brown); Silvio Savarese (U. of Michigan); Octavia Camps (Northeastern); including their own slides. Comparison between Motion Analysis
More informationComputer Vision cmput 428/615
Computer Vision cmput 428/615 Basic 2D and 3D geometry and Camera models Martin Jagersand The equation of projection Intuitively: How do we develop a consistent mathematical framework for projection calculations?
More informationCSE 252B: Computer Vision II
CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribe: Sameer Agarwal LECTURE 1 Image Formation 1.1. The geometry of image formation We begin by considering the process of image formation when a
More informationRange Sensors (time of flight) (1)
Range Sensors (time of flight) (1) Large range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic sensors, infra-red sensors
More informationCS4733 Class Notes, Computer Vision
CS4733 Class Notes, Computer Vision Sources for online computer vision tutorials and demos - http://www.dai.ed.ac.uk/hipr and Computer Vision resources online - http://www.dai.ed.ac.uk/cvonline Vision
More informationRequirements for region detection
Region detectors Requirements for region detection For region detection invariance transformations that should be considered are illumination changes, translation, rotation, scale and full affine transform
More informationModule 5: Video Modeling Lecture 28: Illumination model. The Lecture Contains: Diffuse and Specular Reflection. Objectives_template
The Lecture Contains: Diffuse and Specular Reflection file:///d /...0(Ganesh%20Rana)/MY%20COURSE_Ganesh%20Rana/Prof.%20Sumana%20Gupta/FINAL%20DVSP/lecture%2028/28_1.htm[12/30/2015 4:22:29 PM] Diffuse and
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 21: Light, reflectance and photometric stereo Announcements Final projects Midterm reports due November 24 (next Tuesday) by 11:59pm (upload to CMS) State the
More informationLecture 16: Computer Vision
CS4442/9542b: Artificial Intelligence II Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field
More informationLecture 16: Computer Vision
CS442/542b: Artificial ntelligence Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field Methods
More informationColor and Shading. Color. Shapiro and Stockman, Chapter 6. Color and Machine Vision. Color and Perception
Color and Shading Color Shapiro and Stockman, Chapter 6 Color is an important factor for for human perception for object and material identification, even time of day. Color perception depends upon both
More informationDigital Image Processing. Prof. P.K. Biswas. Department of Electronics & Electrical Communication Engineering
Digital Image Processing Prof. P.K. Biswas Department of Electronics & Electrical Communication Engineering Indian Institute of Technology, Kharagpur Image Segmentation - III Lecture - 31 Hello, welcome
More informationLecture 22: Basic Image Formation CAP 5415
Lecture 22: Basic Image Formation CAP 5415 Today We've talked about the geometry of scenes and how that affects the image We haven't talked about light yet Today, we will talk about image formation and
More informationSchedule for Rest of Semester
Schedule for Rest of Semester Date Lecture Topic 11/20 24 Texture 11/27 25 Review of Statistics & Linear Algebra, Eigenvectors 11/29 26 Eigenvector expansions, Pattern Recognition 12/4 27 Cameras & calibration
More informationFinal Exam Study Guide CSE/EE 486 Fall 2007
Final Exam Study Guide CSE/EE 486 Fall 2007 Lecture 2 Intensity Sufaces and Gradients Image visualized as surface. Terrain concepts. Gradient of functions in 1D and 2D Numerical derivatives. Taylor series.
More informationLecture 8 Object Descriptors
Lecture 8 Object Descriptors Azadeh Fakhrzadeh Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University 2 Reading instructions Chapter 11.1 11.4 in G-W Azadeh Fakhrzadeh
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem
More informationApplication questions. Theoretical questions
The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list
More informationComputer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.
Computer Vision Coordinates Prof. Flávio Cardeal DECOM / CEFET- MG cardeal@decom.cefetmg.br Abstract This lecture discusses world coordinates and homogeneous coordinates, as well as provides an overview
More informationDepth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth
Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze
More informationComputer Vision I - Filtering and Feature detection
Computer Vision I - Filtering and Feature detection Carsten Rother 30/10/2015 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image
More information(Sample) Final Exam with brief answers
Name: Perm #: (Sample) Final Exam with brief answers CS/ECE 181B Intro to Computer Vision March 24, 2017 noon 3:00 pm This is a closed-book test. There are also a few pages of equations, etc. included
More informationLast update: May 4, Vision. CMSC 421: Chapter 24. CMSC 421: Chapter 24 1
Last update: May 4, 200 Vision CMSC 42: Chapter 24 CMSC 42: Chapter 24 Outline Perception generally Image formation Early vision 2D D Object recognition CMSC 42: Chapter 24 2 Perception generally Stimulus
More informationRegion-based Segmentation
Region-based Segmentation Image Segmentation Group similar components (such as, pixels in an image, image frames in a video) to obtain a compact representation. Applications: Finding tumors, veins, etc.
More informationCS231A Midterm Review. Friday 5/6/2016
CS231A Midterm Review Friday 5/6/2016 Outline General Logistics Camera Models Non-perspective cameras Calibration Single View Metrology Epipolar Geometry Structure from Motion Active Stereo and Volumetric
More informationImage processing and features
Image processing and features Gabriele Bleser gabriele.bleser@dfki.de Thanks to Harald Wuest, Folker Wientapper and Marc Pollefeys Introduction Previous lectures: geometry Pose estimation Epipolar geometry
More informationDigital Image Processing
Digital Image Processing Part 9: Representation and Description AASS Learning Systems Lab, Dep. Teknik Room T1209 (Fr, 11-12 o'clock) achim.lilienthal@oru.se Course Book Chapter 11 2011-05-17 Contents
More informationComputer Vision I - Appearance-based Matching and Projective Geometry
Computer Vision I - Appearance-based Matching and Projective Geometry Carsten Rother 05/11/2015 Computer Vision I: Image Formation Process Roadmap for next four lectures Computer Vision I: Image Formation
More informationStereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz
Stereo CSE 576 Ali Farhadi Several slides from Larry Zitnick and Steve Seitz Why do we perceive depth? What do humans use as depth cues? Motion Convergence When watching an object close to us, our eyes
More informationAn introduction to 3D image reconstruction and understanding concepts and ideas
Introduction to 3D image reconstruction An introduction to 3D image reconstruction and understanding concepts and ideas Samuele Carli Martin Hellmich 5 febbraio 2013 1 icsc2013 Carli S. Hellmich M. (CERN)
More informationColour Reading: Chapter 6. Black body radiators
Colour Reading: Chapter 6 Light is produced in different amounts at different wavelengths by each light source Light is differentially reflected at each wavelength, which gives objects their natural colours
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 20: Light, reflectance and photometric stereo Light by Ted Adelson Readings Szeliski, 2.2, 2.3.2 Light by Ted Adelson Readings Szeliski, 2.2, 2.3.2 Properties
More informationLecture 14: Basic Multi-View Geometry
Lecture 14: Basic Multi-View Geometry Stereo If I needed to find out how far point is away from me, I could use triangulation and two views scene point image plane optical center (Graphic from Khurram
More informationGeneralized Hough Transform, line fitting
Generalized Hough Transform, line fitting Introduction to Computer Vision CSE 152 Lecture 11-a Announcements Assignment 2: Due today Midterm: Thursday, May 10 in class Non-maximum suppression For every
More informationLecture: Edge Detection
CMPUT 299 Winter 2007 Lecture: Edge Detection Irene Cheng Overview. What is a pixel in an image? 2. How does Photoshop, + human assistance, detect an edge in a picture/photograph? 3. Behind Photoshop -
More informationMachine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy
1 Machine vision Summary # 11: Stereo vision and epipolar geometry STEREO VISION The goal of stereo vision is to use two cameras to capture 3D scenes. There are two important problems in stereo vision:
More informationVisual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The
More informationComputer Vision I - Basics of Image Processing Part 2
Computer Vision I - Basics of Image Processing Part 2 Carsten Rother 07/11/2014 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image
More informationDigital Image Processing COSC 6380/4393
Digital Image Processing COSC 6380/4393 Lecture 21 Nov 16 th, 2017 Pranav Mantini Ack: Shah. M Image Processing Geometric Transformation Point Operations Filtering (spatial, Frequency) Input Restoration/
More information10/5/09 1. d = 2. Range Sensors (time of flight) (2) Ultrasonic Sensor (time of flight, sound) (1) Ultrasonic Sensor (time of flight, sound) (2) 4.1.
Range Sensors (time of flight) (1) Range Sensors (time of flight) (2) arge range distance measurement -> called range sensors Range information: key element for localization and environment modeling Ultrasonic
More informationECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt
ECE 47: Homework 5 Due Tuesday, October 7 in class @:3pm Seth Hutchinson Luke A Wendt ECE 47 : Homework 5 Consider a camera with focal length λ = Suppose the optical axis of the camera is aligned with
More informationAnnouncements. Hough Transform [ Patented 1962 ] Generalized Hough Transform, line fitting. Assignment 2: Due today Midterm: Thursday, May 5 in class
Announcements Generalized Hough Transform, line fitting Assignment 2: Due today Midterm: Thursday, May 5 in class Introduction to Computer Vision CSE 152 Lecture 11a What is region like if: 1. λ 1 = 0?
More informationChapter 32 Light: Reflection and Refraction. Copyright 2009 Pearson Education, Inc.
Chapter 32 Light: Reflection and Refraction Units of Chapter 32 The Ray Model of Light Reflection; Image Formation by a Plane Mirror Formation of Images by Spherical Mirrors Index of Refraction Refraction:
More informationAnnouncements. Motion. Structure-from-Motion (SFM) Motion. Discrete Motion: Some Counting
Announcements Motion Introduction to Computer Vision CSE 152 Lecture 20 HW 4 due Friday at Midnight Final Exam: Tuesday, 6/12 at 8:00AM-11:00AM, regular classroom Extra Office Hours: Monday 6/11 9:00AM-10:00AM
More informationPerception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich.
Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Perception Sensors Vision Uncertainties, Line extraction
More informationBIL Computer Vision Apr 16, 2014
BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm
More informationAugmented Reality II - Camera Calibration - Gudrun Klinker May 11, 2004
Augmented Reality II - Camera Calibration - Gudrun Klinker May, 24 Literature Richard Hartley and Andrew Zisserman, Multiple View Geometry in Computer Vision, Cambridge University Press, 2. (Section 5,
More informationPhysics for Scientists & Engineers 2
Geometric Optics Physics for Scientists & Engineers 2 Spring Semester 2005 Lecture 36! The study of light divides itself into three fields geometric optics wave optics quantum optics! In the previous chapter,
More information3D Modeling using multiple images Exam January 2008
3D Modeling using multiple images Exam January 2008 All documents are allowed. Answers should be justified. The different sections below are independant. 1 3D Reconstruction A Robust Approche Consider
More informationProjective geometry for Computer Vision
Department of Computer Science and Engineering IIT Delhi NIT, Rourkela March 27, 2010 Overview Pin-hole camera Why projective geometry? Reconstruction Computer vision geometry: main problems Correspondence
More informationTopics and things to know about them:
Practice Final CMSC 427 Distributed Tuesday, December 11, 2007 Review Session, Monday, December 17, 5:00pm, 4424 AV Williams Final: 10:30 AM Wednesday, December 19, 2007 General Guidelines: The final will
More informationComputer Vision Projective Geometry and Calibration. Pinhole cameras
Computer Vision Projective Geometry and Calibration Professor Hager http://www.cs.jhu.edu/~hager Jason Corso http://www.cs.jhu.edu/~jcorso. Pinhole cameras Abstract camera model - box with a small hole
More informationOptics II. Reflection and Mirrors
Optics II Reflection and Mirrors Geometric Optics Using a Ray Approximation Light travels in a straight-line path in a homogeneous medium until it encounters a boundary between two different media The
More informationFeature descriptors and matching
Feature descriptors and matching Detections at multiple scales Invariance of MOPS Intensity Scale Rotation Color and Lighting Out-of-plane rotation Out-of-plane rotation Better representation than color:
More informationLecture 4: Image Processing
Lecture 4: Image Processing Definitions Many graphics techniques that operate only on images Image processing: operations that take images as input, produce images as output In its most general form, an
More informationPart 3: Image Processing
Part 3: Image Processing Image Filtering and Segmentation Georgy Gimel farb COMPSCI 373 Computer Graphics and Image Processing 1 / 60 1 Image filtering 2 Median filtering 3 Mean filtering 4 Image segmentation
More informationStereo Wrap + Motion. Computer Vision I. CSE252A Lecture 17
Stereo Wrap + Motion CSE252A Lecture 17 Some Issues Ambiguity Window size Window shape Lighting Half occluded regions Problem of Occlusion Stereo Constraints CONSTRAINT BRIEF DESCRIPTION 1-D Epipolar Search
More informationPractice Exam Sample Solutions
CS 675 Computer Vision Instructor: Marc Pomplun Practice Exam Sample Solutions Note that in the actual exam, no calculators, no books, and no notes allowed. Question 1: out of points Question 2: out of
More informationOutline 7/2/201011/6/
Outline Pattern recognition in computer vision Background on the development of SIFT SIFT algorithm and some of its variations Computational considerations (SURF) Potential improvement Summary 01 2 Pattern
More informationOther approaches to obtaining 3D structure
Other approaches to obtaining 3D structure Active stereo with structured light Project structured light patterns onto the object simplifies the correspondence problem Allows us to use only one camera camera
More informationAll human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them.
All human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them. - Aristotle University of Texas at Arlington Introduction
More informationComputer Vision Project-1
University of Utah, School Of Computing Computer Vision Project- Singla, Sumedha sumedha.singla@utah.edu (00877456 February, 205 Theoretical Problems. Pinhole Camera (a A straight line in the world space
More informationIntroduction to Computer Vision
Introduction to Computer Vision Michael J. Black Nov 2009 Perspective projection and affine motion Goals Today Perspective projection 3D motion Wed Projects Friday Regularization and robust statistics
More information