Robot Sensing & Perception. METR 4202: Advanced Control & Robotics. Date Lecture (W: 11:10-12:40, ) 30-Jul Introduction 6-Aug

Size: px
Start display at page:

Download "Robot Sensing & Perception. METR 4202: Advanced Control & Robotics. Date Lecture (W: 11:10-12:40, ) 30-Jul Introduction 6-Aug"

Transcription

1 Robot Sensing & Perception Seeing is forgetting the name of what one sees L. Weschler METR 4202: Advanced Control & Robotics Dr Sura Singh -- Lecture # 6 September metr4202@itee.uq.edu.au School of Information Technolog and Electrical Engineering at the Universit of Queensland Schedule Date Lecture W: 11:10-12: Jul Introduction 6-Aug Representing Position & Orientation & State Frames Transformation Matrices & Affine Transformations 13-Aug Robot Kinematics & Ekka Da 20-Aug Robot Dnamics & Control 27-Aug Robot Motion 3-Sep Sensing & Perception 10-Sep Multiple View Geometr Computer Vision 17-Sep Navigation & Localization + Prof. M. Srinivasan 24-Sep Motion Planning + Control 1-Oct Stud break 8-Oct State-Space Modelling 15-Oct Shaping the Dnamic Response 22-Oct Linear Observers & LQR 29-Oct Applications in Industr & Course Review METR 4202: Robotics 3 September

2 Reference Material UQ Librar epdf UQ Librar Hardcop UQ Librar/SpringerLink METR 4202: Robotics 3 September Quick Outline 1. Perception Camera Sensors 1. Image Formation Computational Photograph 2. Calibration 3. Feature etraction 4. Stereopsis and depth 5. Optical flow METR 4202: Robotics 3 September

3 Sensor Information Laser Vision/Cameras GPS METR 4202: Robotics 3 September Mapping: Indoor robots METR 4202: Robotics 3 September

4 Cameras Wikipedia E-30-Cutmodel METR 4202: Robotics 3 September Perception Making Sense from Sensors METR 4202: Robotics 3 September

5 Perception Perception is about understanding the image for informing latter robot / control action METR 4202: Robotics 3 September Perception Perception is about understanding the image for informing latter robot / control action METR 4202: Robotics 3 September

6 Image Formation: Lens Optics Sec. 2.2 from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Image Formation: Lens Optics Chromatic Aberration & Vignetting Chromatic Aberration: Vignetting: Sec. 2.2 from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

7 Image Formation: Lens Optics Aperture / Depth of Field METR 4202: Robotics 3 September Image Formation Single View Geometr Corke Ch. 11 METR 4202: Robotics 3 September

8 Image Formation Single View Geometr [II] Hartl & Zisserman Ch. 6 METR 4202: Robotics 3 September Image Formation Single View Geometr Camera Projection Matri = Image point X = World point K = Camera Calibration Matri Perspective Camera as: where: P is 3 4 and of rank 3 METR 4202: Robotics 3 September

9 Image Formation Two-View Geometr [Stereopsis] Fundamental Matri METR 4202: Robotics 3 September Planar Projective Transformations Perspective projection of a plane lots of names for this: homograph colineation planar projective map Easil modeled using homogeneous coordinates 000 z sss 1 image plane To appl a homograph H s' * * s' * * s * * compute p = Hp p = p /s normalize b dividing b third component Slide from Szeliski Computer Vision: Algorithms and Applications p H * * * 1 METR 4202: Robotics 3 September p 9

10 Transformations Forward Warp Inverse Warp Sec. 3.6 from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Transformations : New Image & : Old Image Euclidean: Distances preserved Similarit Scaled Rotation: Angles preserved Fig. 2.4 from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

11 Transformations [2] Affine : lines remain Projective: straight lines preserved H: Homogenous 33 Matri Fig. 2.4 from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September D Transformations = point in the new or 2 nd image = point in the old image Translation = + t Rotation = R + t Similarit = sr + t Affine = A Projective = A here is an inhomogeneous pt 2-vector is a homogeneous point METR 4202: Robotics 3 September

12 2-D Transformations METR 4202: Robotics 3 September D Transformations Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

13 Image Rectification To unwarp rectif an image solve for H given p and p solve equations of the form: sp = Hp linear in unknowns: s and coefficients of H need at least 4 points Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September D Projective Geometr These concepts generalize naturall to 3D Homogeneous coordinates Projective 3D points have four coords: P = XYZW Dualit A plane L is also represented b a 4-vector Points and planes are dual in 3D: L P=0 Projective transformations Represented b 44 matrices T: P = TP L = L T-1 Lines are a special case Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

14 3D 2D Perspective Projection Image Formation Equations X c Y c Z c f u c u Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September D 2D Perspective Projection Matri Projection camera matri: s * X * Y Z * 1 p s * * * * ΠP s * * * * * It s useful to decompose into T R project A Π s 0 t s t f R 0 I T intrinsics projection orientation position Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

15 Calibration matri Is this form of K good enough? non-square piels digital video skew radial distortion From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Calibration See: Camera Calibration Toolbo for Matlab Intrinsic: Internal Parameters Focal length: The focal length in piels. Principal point: The principal point Skew coefficient: The skew coefficient defining the angle between the and piel aes. Distortions: The image distortion coefficients radial and tangential distortions tpicall two quadratic functions Etrinsics: Where the Camera image plane is placed: Rotations: A set of 33 rotation matrices for each image Translations: A set of 31 translation vectors for each image METR 4202: Robotics 3 September

16 Camera calibration Determine camera parameters from known 3D points or calibration objects internal or intrinsic parameters such as focal length optical center aspect ratio: what kind of camera? eternal or etrinsic pose parameters: where is the camera? How can we do this? From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Camera calibration approaches Possible approaches: linear regression least squares non-linear optimization vanishing points multiple planar patterns panoramas rotational motion From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

17 17 Measurements on Planes You can not just add a tape measure! Approach: unwarp then measure Slide from Szeliski Computer Vision: Algorithms and Applications 3 September METR 4202: Robotics 33 Projection Models Orthographic Weak Perspective Affine Perspective Projective * * * * * * * * Π z z t j j j t i i i Π t R Π * * * * * * * * * * * * Π z z t j j j t i i i f Π Slide from Szeliski Computer Vision: Algorithms and Applications 3 September METR 4202: Robotics 34

18 Properties of Projection Preserves Lines and conics Incidence Invariants cross-ratio Does not preserve Lengths Angles Parallelism METR 4202: Robotics 3 September The Projective Plane Wh do we need homogeneous coordinates? Represent points at infinit homographies perspective projection multi-view relationships What is the geometric intuition? A point in the image is a ra in projective space sss z image plane Each point on the plane is represented b a ra sss all points on the ra are equivalent: 1 s s s Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

19 Projective Lines What is a line in projective space? A line is a plane of ras through origin all ras z satisfing: a + b + cz = 0 in vector notation : 0 a b c z l T p A line is represented as a homogeneous 3-vector l Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Point and Line Dualit A line l is a homogeneous 3-vector a ra It is to ever point ra p on the line: lt p=0 l p 1 p 2 l 1 l 2 What is the line l spanned b ras p 1 and p 2? l is to p 1 and p 2 l = p 1 p 2 l is the plane normal What is the intersection of two lines l 1 and l 2? p is to l 1 and l 2 p = l 1 l 2 Points and lines are dual in projective space ever propert of points also applies to lines p METR 4202: Robotics 3 September

20 Ideal points and lines Ideal point point at infinit p 0 parallel to image plane It has infinite image coordinates ss0 ss0 z image plane z image plane Line at infinit l parallel to image plane Contains all ideal points METR 4202: Robotics 3 September Vanishing Points Vanishing point projection of a point at infinit whiteboard image plane capture architecture vanishing point camera center ground plane Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

21 Fun With Vanishing Points Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Vanishing Points 2D image plane vanishing point camera center line on ground plane METR 4202: Robotics 3 September

22 Vanishing Points Properties An two parallel lines have the same vanishing point The ra from C through v point is parallel to the lines An image ma have more than one vanishing point image plane vanishing point V camera center C line on ground plane line on ground plane METR 4202: Robotics 3 September Vanishing Lines Multiple Vanishing Points An set of parallel lines on the plane define a vanishing point The union of all of these vanishing points is the horizon line v 1 v 2 METR 4202: Robotics 3 September

23 Two-View Geometr: Epipolar Plane Epipole: The point of intersection of the line joining the camera centres the baseline with the image plane. Equivalentl the epipole is the image in one view of the camera centre of the other view. Epipolar plane is a plane containing the baseline. There is a one-parameter famil a pencil of epipolar planes Epipolar line is the intersection of an epipolar plane with the image plane. All epipolar lines intersect at the epipole. An epipolar plane intersects the left and right image planes in epipolar lines and defines the correspondence between the lines. METR 4202: Robotics 3 September Two-frame methods Two main variants: Calibrated: Essential matri E use ra directions i i Uncalibrated: Fundamental matri F [Hartle & Zisserman 2000] From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

24 Fundamental matri Camera calibrations are unknown F = 0 with F = [e] H = K [t] R K-1 Solve for F using least squares SVD re-scale i i so that i 1/2 [Hartle] e epipole is still the least singular vector of F H obtained from the other two s.v.s plane + paralla projective reconstruction use self-calibration to determine K [Pollefes] From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Essential matri Co-planarit constraint: R + t [t] [t] R [t] [t] R E = 0 with E =[t] R Solve for E using least squares SVD t is the least singular vector of E R obtained from the other two s.v.s From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

25 Stereo: Epipolar geometr Match features along epipolar lines epipolar line epipolar plane viewing ra Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Stereo: epipolar geometr for two images or images with collinear camera centers can find epipolar lines epipolar lines are the projection of the pencil of planes passing through the centers Rectification: warping the input images perspective transformation so that epipolar lines are horizontal Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

26 Fundamental Matri The fundamental matri is the algebraic representation of epipolar geometr. METR 4202: Robotics 3 September Fundamental Matri Eample Suppose the camera matrices are those of a calibrated stereo rig with the world origin at the first camera Then: Epipoles are at: METR 4202: Robotics 3 September

27 Summar of fundamental matri properties METR 4202: Robotics 3 September Fundamental Matri & Motion Under a pure translational camera motion 3D points appear to slide along parallel rails. The images of these parallel lines intersect in a vanishing point corresponding to the translation direction. The epipole e is the vanishing point. METR 4202: Robotics 3 September

28 Cool Robotics Share D. Wedge The Fundamental Matri Song METR 4202: Robotics 3 September Rectification Project each image onto same plane which is parallel to the epipole Resample lines and shear/stretch to place lines in correspondence and minimize distortion [Zhang and Loop MSR-TR-99-21] Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

29 Rectification BAD! Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Rectification GOOD! Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

30 Matching criteria Raw piel values correlation Band-pass filtered images [Jones & Malik 92] Corner like features [Zhang ] Edges [man people ] Gradients [Seitz 89; Scharstein 94] Rank statistics [Zabih & Woodfill 94] Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Finding correspondences Appl feature matching criterion e.g. correlation or Lucas-Kanade at all piels simultaneousl Search onl over epipolar lines man fewer candidate positions Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

31 Image registration revisited How do we determine correspondences? block matching or SSD sum squared differences d is the disparit horizontal motion How big should the neighborhood be? Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Neighborhood size Smaller neighborhood: more details Larger neighborhood: fewer isolated mistakes w = 3 w = 20 Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

32 Stereo: certaint modeling Compute certaint map from correlations input depth map certaint map METR 4202: Robotics 3 September Plane Sweep Stereo Sweep famil of planes through volume projective re-sampling of XYZ input image composite virtual camera each plane defines an image composite homograph METR 4202: Robotics 3 September

33 Plane sweep stereo Re-order piel / disparit evaluation loops for ever piel for ever disparit compute cost for ever disparit for ever piel compute cost METR 4202: Robotics 3 September Stereo matching framework For ever disparit compute raw matching costs Wh use a robust function? occlusions other outliers Can also use alternative match criteria METR 4202: Robotics 3 September

34 Stereo matching framework Aggregate costs spatiall Here we are using a bo filter efficient moving average implementation Can also use weighted average [non-linear] diffusion METR 4202: Robotics 3 September Stereo matching framework Choose winning disparit at each piel Interpolate to sub-piel accurac Ed d* d METR 4202: Robotics 3 September

35 Traditional Stereo Matching Advantages: gives detailed surface estimates fast algorithms based on moving averages sub-piel disparit estimates and confidence Limitations: narrow baseline nois estimates fails in tetureless areas gets confused near occlusion boundaries METR 4202: Robotics 3 September Stereo with Non-Linear Diffusion Problem with traditional approach: gets confused near discontinuities New approach: use iterative non-linear aggregation to obtain better estimate provabl equivalent to mean-field estimate of Markov Random Field METR 4202: Robotics 3 September

36 How to get Matching Points? Features Colour Corners Edges Lines Statistics on Edges: SIFT SURF ORB In OpenCV: The following detector tpes are supported: "FAST" FastFeatureDetector "STAR" StarFeatureDetector "SIFT" SIFT nonfree module "SURF" SURF nonfree module "ORB" ORB "BRISK" BRISK "MSER" MSER "GFTT" GoodFeaturesToTrackDetector "HARRIS" GoodFeaturesToTrackDetector with Harris detector enabled "Dense" DenseFeatureDetector "SimpleBlob" SimpleBlobDetector METR 4202: Robotics 3 September Feature-based stereo Match corner interest points Interpolate complete solution Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

37 Features -- Colour Features Fig: Ch. 10 Robotics Vision and Control Baer Patterns RGB is NOT an absolute metric colour space Also! RGB displa or additive colour does not map to CYMK printing or subtractive colour without calibration Y-Cr-Cb or HSV does not solve this either METR 4202: Robotics 3 September How to get the Features? Still MANY Was Cann edge detector: METR 4202: Robotics 3 September

38 Hough Transform Uses a voting mechanism Can be used for other lines and shapes not just straight lines METR 4202: Robotics 3 September Hough Transform: Voting Space Count the number of lines that can go through a point and move it from the - plane to the a-b plane There is onl a one- infinite number a line! of solutions not a two- infinite set a plane METR 4202: Robotics 3 September

39 Hough Transform: Voting Space In practice the polar form is often used This avoids problems with lines that are nearl vertical METR 4202: Robotics 3 September Hough Transform: Algorithm 1. Quantize the parameter space appropriatel. 2. Assume that each cell in the parameter space is an accumulator. Initialize all cells to zero. 3. For each point in the visual & range image space increment b 1 each of the accumulators that satisf the equation. 4. Maima in the accumulator arra correspond to the parameters of model instances. METR 4202: Robotics 3 September

40 Line Detection Hough Lines [1] A line in an image can be epressed as two variables: Cartesian coordinate sstem: mb Polar coordinate sstem: r θ avoids problems with vert. lines =m+b For each point 1 1 we can write: Each pair rθ represents a line that passes through 1 1 See also OpenCV documentation cv::houghlines METR 4202: Robotics 3 September Line Detection Hough Lines [2] Thus a given point gives a sinusoid Repeating for all points on the image See also OpenCV documentation cv::houghlines METR 4202: Robotics 3 September

41 Line Detection Hough Lines [3] Thus a given point gives a sinusoid Repeating for all points on the image NOTE that an intersection of sinusoids represents a point represents a line in which piel points la. Thus a line can be detected b finding the number of Intersections between curves See also OpenCV documentation cv::houghlines METR 4202: Robotics 3 September Cool Robotics Share -- Hough Transform METR 4202: Robotics 3 September

42 Line Etraction and Segmentation Adopted from Williams Fitch and Singh MTRX 4700 METR 4202: Robotics 3 September Line Formula Adopted from Williams Fitch and Singh MTRX 4700 METR 4202: Robotics 3 September

43 Line Estimation Least squares minimization of the line: Line Equation: Error in Fit: Solution: Adopted from Williams Fitch and Singh MTRX 4700 METR 4202: Robotics 3 September Line Splitting / Segmentation What about corners? Split into multiple lines via epectation maimization 1. Epect assume a number of lines N sa 3 2. Find breakpoints b finding nearest neighbours upto a threshold or simpl at random RANSAC 3. How to know N? Also RANSAC Adopted from Williams Fitch and Singh MTRX 4700 METR 4202: Robotics 3 September

44 of a Point from a Line Segment d D Adopted from Williams Fitch and Singh MTRX 4700 METR 4202: Robotics 3 September Edge Detection Cann edge detector: Pepsi Sequence: Image Data: and Szeliski CS223B-L9 See also: Use of Temporal information to aid segmentation: METR 4202: Robotics 3 September

45 Wh etract features? Object detection Robot Navigation Scene Recognition Steps: Etract Features Match Features Adopted drom S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Wh etract features? [2] Panorama stitching Step 3: Align images Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September

46 Characteristics of good features Repeatabilit The same feature can be found in several images despite geometric and photometric transformations Salienc Each feature is distinctive Compactness and efficienc Man fewer features than image piels Localit A feature occupies a relativel small area of the image; robust to clutter and occlusion Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Finding Corners Ke propert: in the region around a corner image gradient has two or more dominant directions Corners are repeatable and distinctive C.Harris and M.Stephens. "A Combined Corner and Edge Detector. Proceedings of the 4th Alve Vision Conference: pages Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September

47 Corner Detection: Basic Idea Look through a window Shifting a window in an direction should give a large change in intensit flat region: no change in all directions Source: A. Efros edge : no change along the edge direction corner : significant change in all directions METR 4202: Robotics 3 September Corner Detection: Mathematics Change in appearance of window w for the shift [uv]: 2 E uv w I u v I I Eu v E32 Adopted from S. Lazebnik Gang Hua CS 558 w METR 4202: Robotics 3 September

48 Corner Detection: Mathematics Change in appearance of window w for the shift [uv]: 2 E uv w I u v I I Eu v E00 Adopted from S. Lazebnik Gang Hua CS 558 w METR 4202: Robotics 3 September Corner Detection: Mathematics Change in appearance of window w for the shift [uv]: 2 E uv w I u v I Window function Shifted intensit Intensit Window function w = or Adopted from S. Lazebnik Gang Hua CS in window 0 outside Gaussian METR 4202: Robotics Source: 3 September R. Szeliski

49 Corner Detection: Mathematics Change in appearance of window w for the shift [uv]: 2 E uv w I u v I We want to find out how this function behaves for small shifts Eu v Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Corner Detection: Mathematics Change in appearance of window w for the shift [uv]: Eu 00 1 Euu 00 E u v E00 [ u v] [ u v] Ev00 2 Euv 00 2 E uv w I u v I We want to find out how this function behaves for small shifts Euv00 u E vv00 v Local quadratic approimation of Euv in the neighborhood of 00 is given b the second-order Talor epansion: Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September

50 50 Corner Detection: Mathematics v u E E E E v u E E v u E v u E vv uv uv uu v u ] [ ] [ 00 2 E uv w I u v I Second-order Talor epansion of Euv about 00: v u I I v u I w v u I v u I w v u E v u I I v u I w v u I v u I w v u E v u I I v u I w v u E uv uu u Adopted from S. Lazebnik Gang Hua CS September METR 4202: Robotics 99 Corner Detection: Mathematics v u I w I I w I I w I w v u v u E 2 2 ] [ 2 E uv w I u v I Second-order Talor epansion of Euv about 00: I I w E I I w E I I w E E E E uv vv uu v u Adopted from S. Lazebnik Gang Hua CS September METR 4202: Robotics 100

51 Harris detector: Steps Compute Gaussian derivatives at each piel Compute second moment matri M in a Gaussian window around each piel Compute corner response function R Threshold R Find local maima of response function nonmaimum suppression C.Harris and M.Stephens. A Combined Corner and Edge Detector. Proceedings of the 4th Alve Vision Conference: pages Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Harris Detector: Steps Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September

52 Harris Detector: Steps Compute corner response R Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Harris Detector: Steps Find points with large corner response: R>threshold Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September

53 Harris Detector: Steps Take onl the points of local maima of R Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Harris Detector: Steps Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September

54 Invariance and covariance We want corner locations to be invariant to photometric transformations and covariant to geometric transformations Invariance: image is transformed and corner locations do not change Covariance: if we have two transformed versions of the same image features should be detected in corresponding locations Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September RANdom SAmple Consensus 1. Repeatedl select a small minimal subset of correspondences 2. Estimate a solution in this case a the line 3. Count the number of inliers e <Θ for LMS estimate med e 4. Pick the best subset of inliers 5. Find a complete least-squares solution Related to least median squares See also: MAPSAC Maimum A Posteriori SAmple Consensus From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

55 Cool Robotics Share Time! D. Wedge The RANSAC Song METR 4202: Robotics 3 September Scale Invariant Feature Transform Basic idea: Take 1616 square window around detected feature Compute edge orientation angle of the gradient - 90 for each piel Throw out weak edges threshold gradient magnitude Create histogram of surviving edge orientations 0 2 angle histogram Adapted from slide b David Lowe METR 4202: Robotics 3 September

56 SIFT descriptor Full version Divide the 1616 window into a 44 grid of cells 22 case shown below Compute an orientation histogram for each cell 16 cells * 8 orientations = 128 dimensional descriptor Adapted from slide b David Lowe METR 4202: Robotics 3 September Properties of SIFT Etraordinaril robust matching technique Can handle changes in viewpoint Up to about 60 degree out of plane rotation Can handle significant changes in illumination Sometimes even da vs. night below Fast and efficient can run in real time Lots of code available From David Lowe and Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

57 Feature matching Given a feature in I 1 how to find the best match in I 2? 1. Define distance function that compares two descriptors 2. Test all the features in I 2 find the one with min distance From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Feature distance How to define the difference between two features f 1 f 2? Simple approach is SSDf 1 f 2 sum of square differences between entries of the two descriptors can give good scores to ver ambiguous bad matches f 1 f 2 I 1 I 2 From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

58 Feature distance How to define the difference between two features f 1 f 2? Better approach: ratio distance = SSDf 1 f 2 / SSDf 1 f 2 f 2 is best SSD match to f 1 in I 2 f 2 is 2 nd best SSD match to f 1 in I 2 gives small values for ambiguous matches f 1 f f ' 2 2 I 1 I 2 From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Evaluating the results How can we measure the performance of a feature matcher? feature distance From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

59 True/false positives true match 200 false match feature distance The distance threshold affects performance True positives = # of detected matches that are correct Suppose we want to maimize these how to choose threshold? False positives = # of detected matches that are incorrect Suppose we want to minimize these how to choose threshold? From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Levenberg-Marquardt Iterative non-linear least squares [Press 92] Linearize measurement equations Substitute into log-likelihood equation: quadratic cost function in Dm From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

60 Levenberg-Marquardt What if it doesn t converge? Multipl diagonal b 1 + l increase l until it does Halve the step size Dm m favorite Use line search Other ideas? Uncertaint analsis: covariance S = A-1 Is maimum likelihood the best idea? How to start in vicinit of global minimum? From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Camera matri calibration Advantages: ver simple to formulate and solve can recover K [R t] from M using QR decomposition [Golub & VanLoan 96] Disadvantages: doesn't compute internal parameters more unknowns than true degrees of freedom need a separate camera matri for each new view From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

61 Multi-plane calibration Use several images of planar target held at unknown orientations [Zhang 99] Compute plane homographies Solve for K-TK-1 from Hk s 1plane if onl f unknown 2 planes if fucvc unknown 3+ planes for full K Code available from Zhang and OpenCV From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Rotational motion Use pure rotation large scene to estimate f estimate f from pairwise homographies re-estimate f from 360º gap optimize over all {KRj} parameters [Stein 95; Hartle 97; Shum & Szeliski 00; Kang & Weiss 99] f=510 f=468 Most accurate wa to get f short of surveing distant points From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

62 SFM: Structure from Motion & Cool Robotics Share this week METR 4202: Robotics 3 September Structure [from] Motion Given a set of feature tracks estimate the 3D structure and 3D camera motion. Assumption: orthographic projection Tracks: u fp v fp f: frame p: point Subtract out mean 2D position i f : rotation s p : position From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

63 Structure from motion How man points do we need to match? 2 frames: Rt: 5 dof + 3n point locations 4n point measurements n 5 k frames: 6k n 2kn alwas want to use man more From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Measurement equations Measurement equations u fp = i ft s p i f : rotation s p : position v fp = j ft s p Stack them up W = R S R = i 1 i F j 1 j F T S = s 1 s P From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

64 Factorization W = R 2F3 S 3P SVD W = U Λ V Λ must be rank 3 W = U Λ 1/2 Λ 1/2 V = U V Make R orthogonal R = QU S = Q -1 V i ft Q T Qi f = 1 From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Results Look at paper figures From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

65 Bundle Adjustment What makes this non-linear minimization hard? man more parameters: potentiall slow poorer conditioning high correlation potentiall lots of outliers gauge coordinate freedom From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Lots of parameters: sparsit Onl a few entries in Jacobian are non-zero From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

66 Sparse Cholesk skline First used in finite element analsis Applied to SfM b [Szeliski & Kang 1994] structure motion fill-in From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Conditioning and gauge freedom Poor conditioning: use 2nd order method use Cholesk decomposition Gauge freedom fi certain parameters orientation or zero out last few rows in Cholesk decomposition From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September

67 More Cool Robotics Share! METR 4202: Robotics 3 September Cool Robotics Share IV Source: Youtube: Wired How the Tesla Model S is Made METR 4202: Robotics 3 September

viral infection edition

viral infection edition METR4810 Mechatronics Team Project 2 Paul Pounds 8 April 2014 Beginner s Guide to Image Processing or How to make vision-guided killbots step 1 viral infection edition Paul Pounds 8 April 2014 Universit

More information

Computer Vision Lecture 20

Computer Vision Lecture 20 Computer Vision Lecture 2 Motion and Optical Flow Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de 28.1.216 Man slides adapted from K. Grauman, S. Seitz, R. Szeliski,

More information

CS 2770: Intro to Computer Vision. Multiple Views. Prof. Adriana Kovashka University of Pittsburgh March 14, 2017

CS 2770: Intro to Computer Vision. Multiple Views. Prof. Adriana Kovashka University of Pittsburgh March 14, 2017 CS 277: Intro to Computer Vision Multiple Views Prof. Adriana Kovashka Universit of Pittsburgh March 4, 27 Plan for toda Affine and projective image transformations Homographies and image mosaics Stereo

More information

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability

Midterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability Midterm Wed. Local features: detection and description Monday March 7 Prof. UT Austin Covers material up until 3/1 Solutions to practice eam handed out today Bring a 8.5 11 sheet of notes if you want Review

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 12 130228 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Panoramas, Mosaics, Stitching Two View Geometry

More information

Wikipedia - Mysid

Wikipedia - Mysid Wikipedia - Mysid Erik Brynjolfsson, MIT Filtering Edges Corners Feature points Also called interest points, key points, etc. Often described as local features. Szeliski 4.1 Slides from Rick Szeliski,

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational

More information

CS4670: Computer Vision

CS4670: Computer Vision CS4670: Computer Vision Noah Snavely Lecture 6: Feature matching and alignment Szeliski: Chapter 6.1 Reading Last time: Corners and blobs Scale-space blob detector: Example Feature descriptors We know

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

calibrated coordinates Linear transformation pixel coordinates

calibrated coordinates Linear transformation pixel coordinates 1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial

More information

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263 Index 3D reconstruction, 125 5+1-point algorithm, 284 5-point algorithm, 270 7-point algorithm, 265 8-point algorithm, 263 affine point, 45 affine transformation, 57 affine transformation group, 57 affine

More information

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford Image matching Harder case by Diva Sian by Diva Sian by scgbt by swashford Even harder case Harder still? How the Afghan Girl was Identified by Her Iris Patterns Read the story NASA Mars Rover images Answer

More information

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford

Harder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford Image matching Harder case by Diva Sian by Diva Sian by scgbt by swashford Even harder case Harder still? How the Afghan Girl was Identified by Her Iris Patterns Read the story NASA Mars Rover images Answer

More information

Structure from motion

Structure from motion Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t R 2 3,t 3 Camera 1 Camera

More information

Camera Geometry II. COS 429 Princeton University

Camera Geometry II. COS 429 Princeton University Camera Geometry II COS 429 Princeton University Outline Projective geometry Vanishing points Application: camera calibration Application: single-view metrology Epipolar geometry Application: stereo correspondence

More information

Computing F class 13. Multiple View Geometry. Comp Marc Pollefeys

Computing F class 13. Multiple View Geometry. Comp Marc Pollefeys Computing F class 3 Multiple View Geometr Comp 90-089 Marc Pollefes Multiple View Geometr course schedule (subject to change) Jan. 7, 9 Intro & motivation Projective D Geometr Jan. 4, 6 (no class) Projective

More information

Final Exam Study Guide

Final Exam Study Guide Final Exam Study Guide Exam Window: 28th April, 12:00am EST to 30th April, 11:59pm EST Description As indicated in class the goal of the exam is to encourage you to review the material from the course.

More information

Image matching. Announcements. Harder case. Even harder case. Project 1 Out today Help session at the end of class. by Diva Sian.

Image matching. Announcements. Harder case. Even harder case. Project 1 Out today Help session at the end of class. by Diva Sian. Announcements Project 1 Out today Help session at the end of class Image matching by Diva Sian by swashford Harder case Even harder case How the Afghan Girl was Identified by Her Iris Patterns Read the

More information

EECS 556 Image Processing W 09

EECS 556 Image Processing W 09 EECS 556 Image Processing W 09 Motion estimation Global vs. Local Motion Block Motion Estimation Optical Flow Estimation (normal equation) Man slides of this lecture are courtes of prof Milanfar (UCSC)

More information

Photo by Carl Warner

Photo by Carl Warner Photo b Carl Warner Photo b Carl Warner Photo b Carl Warner Fitting and Alignment Szeliski 6. Computer Vision CS 43, Brown James Has Acknowledgment: Man slides from Derek Hoiem and Grauman&Leibe 2008 AAAI

More information

3D Photography: Epipolar geometry

3D Photography: Epipolar geometry 3D Photograph: Epipolar geometr Kalin Kolev, Marc Pollefes Spring 203 http://cvg.ethz.ch/teaching/203spring/3dphoto/ Schedule (tentative) Feb 8 Feb 25 Mar 4 Mar Mar 8 Mar 25 Apr Apr 8 Apr 5 Apr 22 Apr

More information

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253 Index 3D reconstruction, 123 5+1-point algorithm, 274 5-point algorithm, 260 7-point algorithm, 255 8-point algorithm, 253 affine point, 43 affine transformation, 55 affine transformation group, 55 affine

More information

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 Image Features: Local Descriptors Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 [Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 2/ 58 Local Features Detection: Identify

More information

Application questions. Theoretical questions

Application questions. Theoretical questions The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list

More information

Transformations Between Two Images. Translation Rotation Rigid Similarity (scaled rotation) Affine Projective Pseudo Perspective Bi linear

Transformations Between Two Images. Translation Rotation Rigid Similarity (scaled rotation) Affine Projective Pseudo Perspective Bi linear Transformations etween Two Images Translation Rotation Rigid Similarit (scaled rotation) ffine Projective Pseudo Perspective i linear Fundamental Matri Lecture 13 pplications Stereo Structure from Motion

More information

Fundamental Matrix. Lecture 13

Fundamental Matrix. Lecture 13 Fundamental Matri Lecture 13 Transformations etween Two Images Translation Rotation Rigid Similarit (scaled rotation) ffine Projective Pseudo Perspective i linear pplications Stereo Structure from Motion

More information

Automatic Image Alignment (feature-based)

Automatic Image Alignment (feature-based) Automatic Image Alignment (feature-based) Mike Nese with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2006 Today s lecture Feature

More information

But First: Multi-View Projective Geometry

But First: Multi-View Projective Geometry View Morphing (Seitz & Dyer, SIGGRAPH 96) Virtual Camera Photograph Morphed View View interpolation (ala McMillan) but no depth no camera information Photograph But First: Multi-View Projective Geometry

More information

Epipolar Constraint. Epipolar Lines. Epipolar Geometry. Another look (with math).

Epipolar Constraint. Epipolar Lines. Epipolar Geometry. Another look (with math). Epipolar Constraint Epipolar Lines Potential 3d points Red point - fied => Blue point lies on a line There are 3 degrees of freedom in the position of a point in space; there are four DOF for image points

More information

Warping, Morphing and Mosaics

Warping, Morphing and Mosaics Computational Photograph and Video: Warping, Morphing and Mosaics Prof. Marc Pollefes Dr. Gabriel Brostow Toda s schedule Last week s recap Warping Morphing Mosaics Toda s schedule Last week s recap Warping

More information

Local features and image matching. Prof. Xin Yang HUST

Local features and image matching. Prof. Xin Yang HUST Local features and image matching Prof. Xin Yang HUST Last time RANSAC for robust geometric transformation estimation Translation, Affine, Homography Image warping Given a 2D transformation T and a source

More information

CS231A Midterm Review. Friday 5/6/2016

CS231A Midterm Review. Friday 5/6/2016 CS231A Midterm Review Friday 5/6/2016 Outline General Logistics Camera Models Non-perspective cameras Calibration Single View Metrology Epipolar Geometry Structure from Motion Active Stereo and Volumetric

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography Computational Photography Matthias Zwicker University of Bern Fall 2009 Today From 2D to 3D using multiple views Introduction Geometry of two views Stereo matching Other applications Multiview geometry

More information

Visual motion. Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys

Visual motion. Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys Visual motion Man slides adapted from S. Seitz, R. Szeliski, M. Pollefes Motion and perceptual organization Sometimes, motion is the onl cue Motion and perceptual organization Sometimes, motion is the

More information

Today s lecture. Image Alignment and Stitching. Readings. Motion models

Today s lecture. Image Alignment and Stitching. Readings. Motion models Today s lecture Image Alignment and Stitching Computer Vision CSE576, Spring 2005 Richard Szeliski Image alignment and stitching motion models cylindrical and spherical warping point-based alignment global

More information

Image Features. Work on project 1. All is Vanity, by C. Allan Gilbert,

Image Features. Work on project 1. All is Vanity, by C. Allan Gilbert, Image Features Work on project 1 All is Vanity, by C. Allan Gilbert, 1873-1929 Feature extrac*on: Corners and blobs c Mo*va*on: Automa*c panoramas Credit: Ma9 Brown Why extract features? Mo*va*on: panorama

More information

Local features: detection and description. Local invariant features

Local features: detection and description. Local invariant features Local features: detection and description Local invariant features Detection of interest points Harris corner detection Scale invariant blob detection: LoG Description of local patches SIFT : Histograms

More information

MAN-522: COMPUTER VISION SET-2 Projections and Camera Calibration

MAN-522: COMPUTER VISION SET-2 Projections and Camera Calibration MAN-522: COMPUTER VISION SET-2 Projections and Camera Calibration Image formation How are objects in the world captured in an image? Phsical parameters of image formation Geometric Tpe of projection Camera

More information

Object Recognition with Invariant Features

Object Recognition with Invariant Features Object Recognition with Invariant Features Definition: Identify objects or scenes and determine their pose and model parameters Applications Industrial automation and inspection Mobile robots, toys, user

More information

N-Views (1) Homographies and Projection

N-Views (1) Homographies and Projection CS 4495 Computer Vision N-Views (1) Homographies and Projection Aaron Bobick School of Interactive Computing Administrivia PS 2: Get SDD and Normalized Correlation working for a given windows size say

More information

Local features: detection and description May 12 th, 2015

Local features: detection and description May 12 th, 2015 Local features: detection and description May 12 th, 2015 Yong Jae Lee UC Davis Announcements PS1 grades up on SmartSite PS1 stats: Mean: 83.26 Standard Dev: 28.51 PS2 deadline extended to Saturday, 11:59

More information

Structure from motion

Structure from motion Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t 2 R 3,t 3 Camera 1 Camera

More information

CS231M Mobile Computer Vision Structure from motion

CS231M Mobile Computer Vision Structure from motion CS231M Mobile Computer Vision Structure from motion - Cameras - Epipolar geometry - Structure from motion Pinhole camera Pinhole perspective projection f o f = focal length o = center of the camera z y

More information

Feature extraction: Corners Harris Corners Pkwy, Charlotte, NC

Feature extraction: Corners Harris Corners Pkwy, Charlotte, NC Featre etraction: Corners 9300 Harris Corners Pkw Charlotte NC Wh etract featres? Motiation: panorama stitching We hae two images how do we combine them? Wh etract featres? Motiation: panorama stitching

More information

Computer Vision Lecture 17

Computer Vision Lecture 17 Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 10 130221 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Canny Edge Detector Hough Transform Feature-Based

More information

Computer Vision Lecture 17

Computer Vision Lecture 17 Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week

More information

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10 Structure from Motion CSE 152 Lecture 10 Announcements Homework 3 is due May 9, 11:59 PM Reading: Chapter 8: Structure from Motion Optional: Multiple View Geometry in Computer Vision, 2nd edition, Hartley

More information

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo --

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo -- Computational Optical Imaging - Optique Numerique -- Multiple View Geometry and Stereo -- Winter 2013 Ivo Ihrke with slides by Thorsten Thormaehlen Feature Detection and Matching Wide-Baseline-Matching

More information

Feature descriptors and matching

Feature descriptors and matching Feature descriptors and matching Detections at multiple scales Invariance of MOPS Intensity Scale Rotation Color and Lighting Out-of-plane rotation Out-of-plane rotation Better representation than color:

More information

Mosaics. Today s Readings

Mosaics. Today s Readings Mosaics VR Seattle: http://www.vrseattle.com/ Full screen panoramas (cubic): http://www.panoramas.dk/ Mars: http://www.panoramas.dk/fullscreen3/f2_mars97.html Today s Readings Szeliski and Shum paper (sections

More information

Scale Invariant Feature Transform (SIFT) CS 763 Ajit Rajwade

Scale Invariant Feature Transform (SIFT) CS 763 Ajit Rajwade Scale Invariant Feature Transform (SIFT) CS 763 Ajit Rajwade What is SIFT? It is a technique for detecting salient stable feature points in an image. For ever such point it also provides a set of features

More information

CS 558: Computer Vision 3 rd Set of Notes

CS 558: Computer Vision 3 rd Set of Notes 1 CS 558: Computer Vision 3 rd Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Overview Denoising Based on slides

More information

An edge is not a line... Edge Detection. Finding lines in an image. Finding lines in an image. How can we detect lines?

An edge is not a line... Edge Detection. Finding lines in an image. Finding lines in an image. How can we detect lines? Edge Detection An edge is not a line... original image Cann edge detector Compute image derivatives if gradient magnitude > τ and the value is a local maimum along gradient direction piel is an edge candidate

More information

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration Camera Calibration Jesus J Caban Schedule! Today:! Camera calibration! Wednesday:! Lecture: Motion & Optical Flow! Monday:! Lecture: Medical Imaging! Final presentations:! Nov 29 th : W. Griffin! Dec 1

More information

Epipolar Geometry and Stereo Vision

Epipolar Geometry and Stereo Vision Epipolar Geometry and Stereo Vision Computer Vision Jia-Bin Huang, Virginia Tech Many slides from S. Seitz and D. Hoiem Last class: Image Stitching Two images with rotation/zoom but no translation. X x

More information

Step-by-Step Model Buidling

Step-by-Step Model Buidling Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure

More information

Feature Detectors and Descriptors: Corners, Lines, etc.

Feature Detectors and Descriptors: Corners, Lines, etc. Feature Detectors and Descriptors: Corners, Lines, etc. Edges vs. Corners Edges = maxima in intensity gradient Edges vs. Corners Corners = lots of variation in direction of gradient in a small neighborhood

More information

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Stitching = alignment + blending geometrical registration

More information

Local Features: Detection, Description & Matching

Local Features: Detection, Description & Matching Local Features: Detection, Description & Matching Lecture 08 Computer Vision Material Citations Dr George Stockman Professor Emeritus, Michigan State University Dr David Lowe Professor, University of British

More information

Unit 3 Multiple View Geometry

Unit 3 Multiple View Geometry Unit 3 Multiple View Geometry Relations between images of a scene Recovering the cameras Recovering the scene structure http://www.robots.ox.ac.uk/~vgg/hzbook/hzbook1.html 3D structure from images Recover

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz Epipolar Geometry Prof. D. Stricker With slides from A. Zisserman, S. Lazebnik, Seitz 1 Outline 1. Short introduction: points and lines 2. Two views geometry: Epipolar geometry Relation point/line in two

More information

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few... STEREO VISION The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Bill Freeman and Antonio Torralba (MIT), including their own

More information

Multi-stable Perception. Necker Cube

Multi-stable Perception. Necker Cube Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuyuki Kayahara Multiple view geometry Stereo vision Epipolar geometry Lowe Hartley and Zisserman Depth map extraction Essential matrix

More information

Multi-stable Perception. Necker Cube

Multi-stable Perception. Necker Cube Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuuki Kaahara Fitting and Alignment Computer Vision Szeliski 6.1 James Has Acknowledgment: Man slides from Derek Hoiem, Lana Lazebnik, and

More information

Image Warping. Many slides from Alyosha Efros + Steve Seitz. Photo by Sean Carroll

Image Warping. Many slides from Alyosha Efros + Steve Seitz. Photo by Sean Carroll Image Warping Man slides from Alosha Efros + Steve Seitz Photo b Sean Carroll Morphing Blend from one object to other with a series of local transformations Image Transformations image filtering: change

More information

Edge and corner detection

Edge and corner detection Edge and corner detection Prof. Stricker Doz. G. Bleser Computer Vision: Object and People Tracking Goals Where is the information in an image? How is an object characterized? How can I find measurements

More information

CSE 252B: Computer Vision II

CSE 252B: Computer Vision II CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribes: Jeremy Pollock and Neil Alldrin LECTURE 14 Robust Feature Matching 14.1. Introduction Last lecture we learned how to find interest points

More information

Structure from Motion

Structure from Motion 11/18/11 Structure from Motion Computer Vision CS 143, Brown James Hays Many slides adapted from Derek Hoiem, Lana Lazebnik, Silvio Saverese, Steve Seitz, and Martial Hebert This class: structure from

More information

CS 532: 3D Computer Vision 7 th Set of Notes

CS 532: 3D Computer Vision 7 th Set of Notes 1 CS 532: 3D Computer Vision 7 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Logistics No class on October

More information

Structure from Motion

Structure from Motion /8/ Structure from Motion Computer Vision CS 43, Brown James Hays Many slides adapted from Derek Hoiem, Lana Lazebnik, Silvio Saverese, Steve Seitz, and Martial Hebert This class: structure from motion

More information

Instance-level recognition

Instance-level recognition Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction

More information

Homographies and RANSAC

Homographies and RANSAC Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position

More information

Local invariant features

Local invariant features Local invariant features Tuesday, Oct 28 Kristen Grauman UT-Austin Today Some more Pset 2 results Pset 2 returned, pick up solutions Pset 3 is posted, due 11/11 Local invariant features Detection of interest

More information

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching -- Computational Optical Imaging - Optique Numerique -- Single and Multiple View Geometry, Stereo matching -- Autumn 2015 Ivo Ihrke with slides by Thorsten Thormaehlen Reminder: Feature Detection and Matching

More information

Lucas-Kanade Image Registration Using Camera Parameters

Lucas-Kanade Image Registration Using Camera Parameters Lucas-Kanade Image Registration Using Camera Parameters Sunghyun Cho a, Hojin Cho a, Yu-Wing Tai b, Young Su Moon c, Junguk Cho c, Shihwa Lee c, and Seungyong Lee a a POSTECH, Pohang, Korea b KAIST, Daejeon,

More information

CSE 527: Introduction to Computer Vision

CSE 527: Introduction to Computer Vision CSE 527: Introduction to Computer Vision Week 5 - Class 1: Matching, Stitching, Registration September 26th, 2017 ??? Recap Today Feature Matching Image Alignment Panoramas HW2! Feature Matches Feature

More information

D-Calib: Calibration Software for Multiple Cameras System

D-Calib: Calibration Software for Multiple Cameras System D-Calib: Calibration Software for Multiple Cameras Sstem uko Uematsu Tomoaki Teshima Hideo Saito Keio Universit okohama Japan {u-ko tomoaki saito}@ozawa.ics.keio.ac.jp Cao Honghua Librar Inc. Japan cao@librar-inc.co.jp

More information

Image formation - About the course. Grading & Project. Tentative Schedule. Course Content. Students introduction

Image formation - About the course. Grading & Project. Tentative Schedule. Course Content. Students introduction About the course Instructors: Haibin Ling (hbling@temple, Wachman 305) Hours Lecture: Tuesda 5:30-8:00pm, TTLMAN 403B Office hour: Tuesda 3:00-5:00pm, or b appointment Tetbook Computer Vision: Models,

More information

3D Geometry and Camera Calibration

3D Geometry and Camera Calibration 3D Geometr and Camera Calibration 3D Coordinate Sstems Right-handed vs. left-handed 2D Coordinate Sstems ais up vs. ais down Origin at center vs. corner Will often write (u, v) for image coordinates v

More information

Today s lecture. Structure from Motion. Today s lecture. Parameterizing rotations

Today s lecture. Structure from Motion. Today s lecture. Parameterizing rotations Today s lecture Structure from Motion Computer Vision CSE576, Spring 2008 Richard Szeliski Geometric camera calibration camera matrix (Direct Linear Transform) non-linear least squares separating intrinsics

More information

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA

More information

CS6670: Computer Vision

CS6670: Computer Vision CS6670: Computer Vision Noah Snavely Lecture 7: Image Alignment and Panoramas What s inside your fridge? http://www.cs.washington.edu/education/courses/cse590ss/01wi/ Projection matrix intrinsics projection

More information

Feature Based Registration - Image Alignment

Feature Based Registration - Image Alignment Feature Based Registration - Image Alignment Image Registration Image registration is the process of estimating an optimal transformation between two or more images. Many slides from Alexei Efros http://graphics.cs.cmu.edu/courses/15-463/2007_fall/463.html

More information

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects Intelligent Control Systems Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects Shingo Kagami Graduate School of Information Sciences, Tohoku University swk(at)ic.is.tohoku.ac.jp http://www.ic.is.tohoku.ac.jp/ja/swk/

More information

Automatic Image Alignment

Automatic Image Alignment Automatic Image Alignment with a lot of slides stolen from Steve Seitz and Rick Szeliski Mike Nese CS194: Image Manipulation & Computational Photography Alexei Efros, UC Berkeley, Fall 2018 Live Homography

More information

Hartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne

Hartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne Hartley - Zisserman reading club Part I: Hartley and Zisserman Appendix 6: Iterative estimation methods Part II: Zhengyou Zhang: A Flexible New Technique for Camera Calibration Presented by Daniel Fontijne

More information

Instance-level recognition

Instance-level recognition Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction

More information

CS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka

CS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka CS223b Midterm Exam, Computer Vision Monday February 25th, Winter 2008, Prof. Jana Kosecka Your name email This exam is 8 pages long including cover page. Make sure your exam is not missing any pages.

More information

Geometric camera models and calibration

Geometric camera models and calibration Geometric camera models and calibration http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 13 Course announcements Homework 3 is out. - Due October

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

Rectification and Distortion Correction

Rectification and Distortion Correction Rectification and Distortion Correction Hagen Spies March 12, 2003 Computer Vision Laboratory Department of Electrical Engineering Linköping University, Sweden Contents Distortion Correction Rectification

More information

Geometric Model of Camera

Geometric Model of Camera Geometric Model of Camera Dr. Gerhard Roth COMP 42A Winter 25 Version 2 Similar Triangles 2 Geometric Model of Camera Perspective projection P(X,Y,Z) p(,) f X Z f Y Z 3 Parallel lines aren t 4 Figure b

More information

Final Exam Study Guide CSE/EE 486 Fall 2007

Final Exam Study Guide CSE/EE 486 Fall 2007 Final Exam Study Guide CSE/EE 486 Fall 2007 Lecture 2 Intensity Sufaces and Gradients Image visualized as surface. Terrain concepts. Gradient of functions in 1D and 2D Numerical derivatives. Taylor series.

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction

More information

Motion Estimation and Optical Flow Tracking

Motion Estimation and Optical Flow Tracking Image Matching Image Retrieval Object Recognition Motion Estimation and Optical Flow Tracking Example: Mosiacing (Panorama) M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 Example 3D Reconstruction

More information

Instance-level recognition part 2

Instance-level recognition part 2 Visual Recognition and Machine Learning Summer School Paris 2011 Instance-level recognition part 2 Josef Sivic http://www.di.ens.fr/~josef INRIA, WILLOW, ENS/INRIA/CNRS UMR 8548 Laboratoire d Informatique,

More information