Robot Sensing & Perception. METR 4202: Advanced Control & Robotics. Date Lecture (W: 11:10-12:40, ) 30-Jul Introduction 6-Aug
|
|
- Howard Stokes
- 5 years ago
- Views:
Transcription
1 Robot Sensing & Perception Seeing is forgetting the name of what one sees L. Weschler METR 4202: Advanced Control & Robotics Dr Sura Singh -- Lecture # 6 September metr4202@itee.uq.edu.au School of Information Technolog and Electrical Engineering at the Universit of Queensland Schedule Date Lecture W: 11:10-12: Jul Introduction 6-Aug Representing Position & Orientation & State Frames Transformation Matrices & Affine Transformations 13-Aug Robot Kinematics & Ekka Da 20-Aug Robot Dnamics & Control 27-Aug Robot Motion 3-Sep Sensing & Perception 10-Sep Multiple View Geometr Computer Vision 17-Sep Navigation & Localization + Prof. M. Srinivasan 24-Sep Motion Planning + Control 1-Oct Stud break 8-Oct State-Space Modelling 15-Oct Shaping the Dnamic Response 22-Oct Linear Observers & LQR 29-Oct Applications in Industr & Course Review METR 4202: Robotics 3 September
2 Reference Material UQ Librar epdf UQ Librar Hardcop UQ Librar/SpringerLink METR 4202: Robotics 3 September Quick Outline 1. Perception Camera Sensors 1. Image Formation Computational Photograph 2. Calibration 3. Feature etraction 4. Stereopsis and depth 5. Optical flow METR 4202: Robotics 3 September
3 Sensor Information Laser Vision/Cameras GPS METR 4202: Robotics 3 September Mapping: Indoor robots METR 4202: Robotics 3 September
4 Cameras Wikipedia E-30-Cutmodel METR 4202: Robotics 3 September Perception Making Sense from Sensors METR 4202: Robotics 3 September
5 Perception Perception is about understanding the image for informing latter robot / control action METR 4202: Robotics 3 September Perception Perception is about understanding the image for informing latter robot / control action METR 4202: Robotics 3 September
6 Image Formation: Lens Optics Sec. 2.2 from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Image Formation: Lens Optics Chromatic Aberration & Vignetting Chromatic Aberration: Vignetting: Sec. 2.2 from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
7 Image Formation: Lens Optics Aperture / Depth of Field METR 4202: Robotics 3 September Image Formation Single View Geometr Corke Ch. 11 METR 4202: Robotics 3 September
8 Image Formation Single View Geometr [II] Hartl & Zisserman Ch. 6 METR 4202: Robotics 3 September Image Formation Single View Geometr Camera Projection Matri = Image point X = World point K = Camera Calibration Matri Perspective Camera as: where: P is 3 4 and of rank 3 METR 4202: Robotics 3 September
9 Image Formation Two-View Geometr [Stereopsis] Fundamental Matri METR 4202: Robotics 3 September Planar Projective Transformations Perspective projection of a plane lots of names for this: homograph colineation planar projective map Easil modeled using homogeneous coordinates 000 z sss 1 image plane To appl a homograph H s' * * s' * * s * * compute p = Hp p = p /s normalize b dividing b third component Slide from Szeliski Computer Vision: Algorithms and Applications p H * * * 1 METR 4202: Robotics 3 September p 9
10 Transformations Forward Warp Inverse Warp Sec. 3.6 from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Transformations : New Image & : Old Image Euclidean: Distances preserved Similarit Scaled Rotation: Angles preserved Fig. 2.4 from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
11 Transformations [2] Affine : lines remain Projective: straight lines preserved H: Homogenous 33 Matri Fig. 2.4 from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September D Transformations = point in the new or 2 nd image = point in the old image Translation = + t Rotation = R + t Similarit = sr + t Affine = A Projective = A here is an inhomogeneous pt 2-vector is a homogeneous point METR 4202: Robotics 3 September
12 2-D Transformations METR 4202: Robotics 3 September D Transformations Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
13 Image Rectification To unwarp rectif an image solve for H given p and p solve equations of the form: sp = Hp linear in unknowns: s and coefficients of H need at least 4 points Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September D Projective Geometr These concepts generalize naturall to 3D Homogeneous coordinates Projective 3D points have four coords: P = XYZW Dualit A plane L is also represented b a 4-vector Points and planes are dual in 3D: L P=0 Projective transformations Represented b 44 matrices T: P = TP L = L T-1 Lines are a special case Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
14 3D 2D Perspective Projection Image Formation Equations X c Y c Z c f u c u Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September D 2D Perspective Projection Matri Projection camera matri: s * X * Y Z * 1 p s * * * * ΠP s * * * * * It s useful to decompose into T R project A Π s 0 t s t f R 0 I T intrinsics projection orientation position Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
15 Calibration matri Is this form of K good enough? non-square piels digital video skew radial distortion From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Calibration See: Camera Calibration Toolbo for Matlab Intrinsic: Internal Parameters Focal length: The focal length in piels. Principal point: The principal point Skew coefficient: The skew coefficient defining the angle between the and piel aes. Distortions: The image distortion coefficients radial and tangential distortions tpicall two quadratic functions Etrinsics: Where the Camera image plane is placed: Rotations: A set of 33 rotation matrices for each image Translations: A set of 31 translation vectors for each image METR 4202: Robotics 3 September
16 Camera calibration Determine camera parameters from known 3D points or calibration objects internal or intrinsic parameters such as focal length optical center aspect ratio: what kind of camera? eternal or etrinsic pose parameters: where is the camera? How can we do this? From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Camera calibration approaches Possible approaches: linear regression least squares non-linear optimization vanishing points multiple planar patterns panoramas rotational motion From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
17 17 Measurements on Planes You can not just add a tape measure! Approach: unwarp then measure Slide from Szeliski Computer Vision: Algorithms and Applications 3 September METR 4202: Robotics 33 Projection Models Orthographic Weak Perspective Affine Perspective Projective * * * * * * * * Π z z t j j j t i i i Π t R Π * * * * * * * * * * * * Π z z t j j j t i i i f Π Slide from Szeliski Computer Vision: Algorithms and Applications 3 September METR 4202: Robotics 34
18 Properties of Projection Preserves Lines and conics Incidence Invariants cross-ratio Does not preserve Lengths Angles Parallelism METR 4202: Robotics 3 September The Projective Plane Wh do we need homogeneous coordinates? Represent points at infinit homographies perspective projection multi-view relationships What is the geometric intuition? A point in the image is a ra in projective space sss z image plane Each point on the plane is represented b a ra sss all points on the ra are equivalent: 1 s s s Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
19 Projective Lines What is a line in projective space? A line is a plane of ras through origin all ras z satisfing: a + b + cz = 0 in vector notation : 0 a b c z l T p A line is represented as a homogeneous 3-vector l Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Point and Line Dualit A line l is a homogeneous 3-vector a ra It is to ever point ra p on the line: lt p=0 l p 1 p 2 l 1 l 2 What is the line l spanned b ras p 1 and p 2? l is to p 1 and p 2 l = p 1 p 2 l is the plane normal What is the intersection of two lines l 1 and l 2? p is to l 1 and l 2 p = l 1 l 2 Points and lines are dual in projective space ever propert of points also applies to lines p METR 4202: Robotics 3 September
20 Ideal points and lines Ideal point point at infinit p 0 parallel to image plane It has infinite image coordinates ss0 ss0 z image plane z image plane Line at infinit l parallel to image plane Contains all ideal points METR 4202: Robotics 3 September Vanishing Points Vanishing point projection of a point at infinit whiteboard image plane capture architecture vanishing point camera center ground plane Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
21 Fun With Vanishing Points Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Vanishing Points 2D image plane vanishing point camera center line on ground plane METR 4202: Robotics 3 September
22 Vanishing Points Properties An two parallel lines have the same vanishing point The ra from C through v point is parallel to the lines An image ma have more than one vanishing point image plane vanishing point V camera center C line on ground plane line on ground plane METR 4202: Robotics 3 September Vanishing Lines Multiple Vanishing Points An set of parallel lines on the plane define a vanishing point The union of all of these vanishing points is the horizon line v 1 v 2 METR 4202: Robotics 3 September
23 Two-View Geometr: Epipolar Plane Epipole: The point of intersection of the line joining the camera centres the baseline with the image plane. Equivalentl the epipole is the image in one view of the camera centre of the other view. Epipolar plane is a plane containing the baseline. There is a one-parameter famil a pencil of epipolar planes Epipolar line is the intersection of an epipolar plane with the image plane. All epipolar lines intersect at the epipole. An epipolar plane intersects the left and right image planes in epipolar lines and defines the correspondence between the lines. METR 4202: Robotics 3 September Two-frame methods Two main variants: Calibrated: Essential matri E use ra directions i i Uncalibrated: Fundamental matri F [Hartle & Zisserman 2000] From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
24 Fundamental matri Camera calibrations are unknown F = 0 with F = [e] H = K [t] R K-1 Solve for F using least squares SVD re-scale i i so that i 1/2 [Hartle] e epipole is still the least singular vector of F H obtained from the other two s.v.s plane + paralla projective reconstruction use self-calibration to determine K [Pollefes] From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Essential matri Co-planarit constraint: R + t [t] [t] R [t] [t] R E = 0 with E =[t] R Solve for E using least squares SVD t is the least singular vector of E R obtained from the other two s.v.s From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
25 Stereo: Epipolar geometr Match features along epipolar lines epipolar line epipolar plane viewing ra Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Stereo: epipolar geometr for two images or images with collinear camera centers can find epipolar lines epipolar lines are the projection of the pencil of planes passing through the centers Rectification: warping the input images perspective transformation so that epipolar lines are horizontal Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
26 Fundamental Matri The fundamental matri is the algebraic representation of epipolar geometr. METR 4202: Robotics 3 September Fundamental Matri Eample Suppose the camera matrices are those of a calibrated stereo rig with the world origin at the first camera Then: Epipoles are at: METR 4202: Robotics 3 September
27 Summar of fundamental matri properties METR 4202: Robotics 3 September Fundamental Matri & Motion Under a pure translational camera motion 3D points appear to slide along parallel rails. The images of these parallel lines intersect in a vanishing point corresponding to the translation direction. The epipole e is the vanishing point. METR 4202: Robotics 3 September
28 Cool Robotics Share D. Wedge The Fundamental Matri Song METR 4202: Robotics 3 September Rectification Project each image onto same plane which is parallel to the epipole Resample lines and shear/stretch to place lines in correspondence and minimize distortion [Zhang and Loop MSR-TR-99-21] Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
29 Rectification BAD! Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Rectification GOOD! Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
30 Matching criteria Raw piel values correlation Band-pass filtered images [Jones & Malik 92] Corner like features [Zhang ] Edges [man people ] Gradients [Seitz 89; Scharstein 94] Rank statistics [Zabih & Woodfill 94] Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Finding correspondences Appl feature matching criterion e.g. correlation or Lucas-Kanade at all piels simultaneousl Search onl over epipolar lines man fewer candidate positions Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
31 Image registration revisited How do we determine correspondences? block matching or SSD sum squared differences d is the disparit horizontal motion How big should the neighborhood be? Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Neighborhood size Smaller neighborhood: more details Larger neighborhood: fewer isolated mistakes w = 3 w = 20 Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
32 Stereo: certaint modeling Compute certaint map from correlations input depth map certaint map METR 4202: Robotics 3 September Plane Sweep Stereo Sweep famil of planes through volume projective re-sampling of XYZ input image composite virtual camera each plane defines an image composite homograph METR 4202: Robotics 3 September
33 Plane sweep stereo Re-order piel / disparit evaluation loops for ever piel for ever disparit compute cost for ever disparit for ever piel compute cost METR 4202: Robotics 3 September Stereo matching framework For ever disparit compute raw matching costs Wh use a robust function? occlusions other outliers Can also use alternative match criteria METR 4202: Robotics 3 September
34 Stereo matching framework Aggregate costs spatiall Here we are using a bo filter efficient moving average implementation Can also use weighted average [non-linear] diffusion METR 4202: Robotics 3 September Stereo matching framework Choose winning disparit at each piel Interpolate to sub-piel accurac Ed d* d METR 4202: Robotics 3 September
35 Traditional Stereo Matching Advantages: gives detailed surface estimates fast algorithms based on moving averages sub-piel disparit estimates and confidence Limitations: narrow baseline nois estimates fails in tetureless areas gets confused near occlusion boundaries METR 4202: Robotics 3 September Stereo with Non-Linear Diffusion Problem with traditional approach: gets confused near discontinuities New approach: use iterative non-linear aggregation to obtain better estimate provabl equivalent to mean-field estimate of Markov Random Field METR 4202: Robotics 3 September
36 How to get Matching Points? Features Colour Corners Edges Lines Statistics on Edges: SIFT SURF ORB In OpenCV: The following detector tpes are supported: "FAST" FastFeatureDetector "STAR" StarFeatureDetector "SIFT" SIFT nonfree module "SURF" SURF nonfree module "ORB" ORB "BRISK" BRISK "MSER" MSER "GFTT" GoodFeaturesToTrackDetector "HARRIS" GoodFeaturesToTrackDetector with Harris detector enabled "Dense" DenseFeatureDetector "SimpleBlob" SimpleBlobDetector METR 4202: Robotics 3 September Feature-based stereo Match corner interest points Interpolate complete solution Slide from Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
37 Features -- Colour Features Fig: Ch. 10 Robotics Vision and Control Baer Patterns RGB is NOT an absolute metric colour space Also! RGB displa or additive colour does not map to CYMK printing or subtractive colour without calibration Y-Cr-Cb or HSV does not solve this either METR 4202: Robotics 3 September How to get the Features? Still MANY Was Cann edge detector: METR 4202: Robotics 3 September
38 Hough Transform Uses a voting mechanism Can be used for other lines and shapes not just straight lines METR 4202: Robotics 3 September Hough Transform: Voting Space Count the number of lines that can go through a point and move it from the - plane to the a-b plane There is onl a one- infinite number a line! of solutions not a two- infinite set a plane METR 4202: Robotics 3 September
39 Hough Transform: Voting Space In practice the polar form is often used This avoids problems with lines that are nearl vertical METR 4202: Robotics 3 September Hough Transform: Algorithm 1. Quantize the parameter space appropriatel. 2. Assume that each cell in the parameter space is an accumulator. Initialize all cells to zero. 3. For each point in the visual & range image space increment b 1 each of the accumulators that satisf the equation. 4. Maima in the accumulator arra correspond to the parameters of model instances. METR 4202: Robotics 3 September
40 Line Detection Hough Lines [1] A line in an image can be epressed as two variables: Cartesian coordinate sstem: mb Polar coordinate sstem: r θ avoids problems with vert. lines =m+b For each point 1 1 we can write: Each pair rθ represents a line that passes through 1 1 See also OpenCV documentation cv::houghlines METR 4202: Robotics 3 September Line Detection Hough Lines [2] Thus a given point gives a sinusoid Repeating for all points on the image See also OpenCV documentation cv::houghlines METR 4202: Robotics 3 September
41 Line Detection Hough Lines [3] Thus a given point gives a sinusoid Repeating for all points on the image NOTE that an intersection of sinusoids represents a point represents a line in which piel points la. Thus a line can be detected b finding the number of Intersections between curves See also OpenCV documentation cv::houghlines METR 4202: Robotics 3 September Cool Robotics Share -- Hough Transform METR 4202: Robotics 3 September
42 Line Etraction and Segmentation Adopted from Williams Fitch and Singh MTRX 4700 METR 4202: Robotics 3 September Line Formula Adopted from Williams Fitch and Singh MTRX 4700 METR 4202: Robotics 3 September
43 Line Estimation Least squares minimization of the line: Line Equation: Error in Fit: Solution: Adopted from Williams Fitch and Singh MTRX 4700 METR 4202: Robotics 3 September Line Splitting / Segmentation What about corners? Split into multiple lines via epectation maimization 1. Epect assume a number of lines N sa 3 2. Find breakpoints b finding nearest neighbours upto a threshold or simpl at random RANSAC 3. How to know N? Also RANSAC Adopted from Williams Fitch and Singh MTRX 4700 METR 4202: Robotics 3 September
44 of a Point from a Line Segment d D Adopted from Williams Fitch and Singh MTRX 4700 METR 4202: Robotics 3 September Edge Detection Cann edge detector: Pepsi Sequence: Image Data: and Szeliski CS223B-L9 See also: Use of Temporal information to aid segmentation: METR 4202: Robotics 3 September
45 Wh etract features? Object detection Robot Navigation Scene Recognition Steps: Etract Features Match Features Adopted drom S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Wh etract features? [2] Panorama stitching Step 3: Align images Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September
46 Characteristics of good features Repeatabilit The same feature can be found in several images despite geometric and photometric transformations Salienc Each feature is distinctive Compactness and efficienc Man fewer features than image piels Localit A feature occupies a relativel small area of the image; robust to clutter and occlusion Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Finding Corners Ke propert: in the region around a corner image gradient has two or more dominant directions Corners are repeatable and distinctive C.Harris and M.Stephens. "A Combined Corner and Edge Detector. Proceedings of the 4th Alve Vision Conference: pages Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September
47 Corner Detection: Basic Idea Look through a window Shifting a window in an direction should give a large change in intensit flat region: no change in all directions Source: A. Efros edge : no change along the edge direction corner : significant change in all directions METR 4202: Robotics 3 September Corner Detection: Mathematics Change in appearance of window w for the shift [uv]: 2 E uv w I u v I I Eu v E32 Adopted from S. Lazebnik Gang Hua CS 558 w METR 4202: Robotics 3 September
48 Corner Detection: Mathematics Change in appearance of window w for the shift [uv]: 2 E uv w I u v I I Eu v E00 Adopted from S. Lazebnik Gang Hua CS 558 w METR 4202: Robotics 3 September Corner Detection: Mathematics Change in appearance of window w for the shift [uv]: 2 E uv w I u v I Window function Shifted intensit Intensit Window function w = or Adopted from S. Lazebnik Gang Hua CS in window 0 outside Gaussian METR 4202: Robotics Source: 3 September R. Szeliski
49 Corner Detection: Mathematics Change in appearance of window w for the shift [uv]: 2 E uv w I u v I We want to find out how this function behaves for small shifts Eu v Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Corner Detection: Mathematics Change in appearance of window w for the shift [uv]: Eu 00 1 Euu 00 E u v E00 [ u v] [ u v] Ev00 2 Euv 00 2 E uv w I u v I We want to find out how this function behaves for small shifts Euv00 u E vv00 v Local quadratic approimation of Euv in the neighborhood of 00 is given b the second-order Talor epansion: Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September
50 50 Corner Detection: Mathematics v u E E E E v u E E v u E v u E vv uv uv uu v u ] [ ] [ 00 2 E uv w I u v I Second-order Talor epansion of Euv about 00: v u I I v u I w v u I v u I w v u E v u I I v u I w v u I v u I w v u E v u I I v u I w v u E uv uu u Adopted from S. Lazebnik Gang Hua CS September METR 4202: Robotics 99 Corner Detection: Mathematics v u I w I I w I I w I w v u v u E 2 2 ] [ 2 E uv w I u v I Second-order Talor epansion of Euv about 00: I I w E I I w E I I w E E E E uv vv uu v u Adopted from S. Lazebnik Gang Hua CS September METR 4202: Robotics 100
51 Harris detector: Steps Compute Gaussian derivatives at each piel Compute second moment matri M in a Gaussian window around each piel Compute corner response function R Threshold R Find local maima of response function nonmaimum suppression C.Harris and M.Stephens. A Combined Corner and Edge Detector. Proceedings of the 4th Alve Vision Conference: pages Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Harris Detector: Steps Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September
52 Harris Detector: Steps Compute corner response R Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Harris Detector: Steps Find points with large corner response: R>threshold Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September
53 Harris Detector: Steps Take onl the points of local maima of R Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September Harris Detector: Steps Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September
54 Invariance and covariance We want corner locations to be invariant to photometric transformations and covariant to geometric transformations Invariance: image is transformed and corner locations do not change Covariance: if we have two transformed versions of the same image features should be detected in corresponding locations Adopted from S. Lazebnik Gang Hua CS 558 METR 4202: Robotics 3 September RANdom SAmple Consensus 1. Repeatedl select a small minimal subset of correspondences 2. Estimate a solution in this case a the line 3. Count the number of inliers e <Θ for LMS estimate med e 4. Pick the best subset of inliers 5. Find a complete least-squares solution Related to least median squares See also: MAPSAC Maimum A Posteriori SAmple Consensus From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
55 Cool Robotics Share Time! D. Wedge The RANSAC Song METR 4202: Robotics 3 September Scale Invariant Feature Transform Basic idea: Take 1616 square window around detected feature Compute edge orientation angle of the gradient - 90 for each piel Throw out weak edges threshold gradient magnitude Create histogram of surviving edge orientations 0 2 angle histogram Adapted from slide b David Lowe METR 4202: Robotics 3 September
56 SIFT descriptor Full version Divide the 1616 window into a 44 grid of cells 22 case shown below Compute an orientation histogram for each cell 16 cells * 8 orientations = 128 dimensional descriptor Adapted from slide b David Lowe METR 4202: Robotics 3 September Properties of SIFT Etraordinaril robust matching technique Can handle changes in viewpoint Up to about 60 degree out of plane rotation Can handle significant changes in illumination Sometimes even da vs. night below Fast and efficient can run in real time Lots of code available From David Lowe and Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
57 Feature matching Given a feature in I 1 how to find the best match in I 2? 1. Define distance function that compares two descriptors 2. Test all the features in I 2 find the one with min distance From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Feature distance How to define the difference between two features f 1 f 2? Simple approach is SSDf 1 f 2 sum of square differences between entries of the two descriptors can give good scores to ver ambiguous bad matches f 1 f 2 I 1 I 2 From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
58 Feature distance How to define the difference between two features f 1 f 2? Better approach: ratio distance = SSDf 1 f 2 / SSDf 1 f 2 f 2 is best SSD match to f 1 in I 2 f 2 is 2 nd best SSD match to f 1 in I 2 gives small values for ambiguous matches f 1 f f ' 2 2 I 1 I 2 From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Evaluating the results How can we measure the performance of a feature matcher? feature distance From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
59 True/false positives true match 200 false match feature distance The distance threshold affects performance True positives = # of detected matches that are correct Suppose we want to maimize these how to choose threshold? False positives = # of detected matches that are incorrect Suppose we want to minimize these how to choose threshold? From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Levenberg-Marquardt Iterative non-linear least squares [Press 92] Linearize measurement equations Substitute into log-likelihood equation: quadratic cost function in Dm From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
60 Levenberg-Marquardt What if it doesn t converge? Multipl diagonal b 1 + l increase l until it does Halve the step size Dm m favorite Use line search Other ideas? Uncertaint analsis: covariance S = A-1 Is maimum likelihood the best idea? How to start in vicinit of global minimum? From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Camera matri calibration Advantages: ver simple to formulate and solve can recover K [R t] from M using QR decomposition [Golub & VanLoan 96] Disadvantages: doesn't compute internal parameters more unknowns than true degrees of freedom need a separate camera matri for each new view From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
61 Multi-plane calibration Use several images of planar target held at unknown orientations [Zhang 99] Compute plane homographies Solve for K-TK-1 from Hk s 1plane if onl f unknown 2 planes if fucvc unknown 3+ planes for full K Code available from Zhang and OpenCV From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Rotational motion Use pure rotation large scene to estimate f estimate f from pairwise homographies re-estimate f from 360º gap optimize over all {KRj} parameters [Stein 95; Hartle 97; Shum & Szeliski 00; Kang & Weiss 99] f=510 f=468 Most accurate wa to get f short of surveing distant points From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
62 SFM: Structure from Motion & Cool Robotics Share this week METR 4202: Robotics 3 September Structure [from] Motion Given a set of feature tracks estimate the 3D structure and 3D camera motion. Assumption: orthographic projection Tracks: u fp v fp f: frame p: point Subtract out mean 2D position i f : rotation s p : position From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
63 Structure from motion How man points do we need to match? 2 frames: Rt: 5 dof + 3n point locations 4n point measurements n 5 k frames: 6k n 2kn alwas want to use man more From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Measurement equations Measurement equations u fp = i ft s p i f : rotation s p : position v fp = j ft s p Stack them up W = R S R = i 1 i F j 1 j F T S = s 1 s P From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
64 Factorization W = R 2F3 S 3P SVD W = U Λ V Λ must be rank 3 W = U Λ 1/2 Λ 1/2 V = U V Make R orthogonal R = QU S = Q -1 V i ft Q T Qi f = 1 From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Results Look at paper figures From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
65 Bundle Adjustment What makes this non-linear minimization hard? man more parameters: potentiall slow poorer conditioning high correlation potentiall lots of outliers gauge coordinate freedom From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Lots of parameters: sparsit Onl a few entries in Jacobian are non-zero From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
66 Sparse Cholesk skline First used in finite element analsis Applied to SfM b [Szeliski & Kang 1994] structure motion fill-in From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September Conditioning and gauge freedom Poor conditioning: use 2nd order method use Cholesk decomposition Gauge freedom fi certain parameters orientation or zero out last few rows in Cholesk decomposition From Szeliski Computer Vision: Algorithms and Applications METR 4202: Robotics 3 September
67 More Cool Robotics Share! METR 4202: Robotics 3 September Cool Robotics Share IV Source: Youtube: Wired How the Tesla Model S is Made METR 4202: Robotics 3 September
viral infection edition
METR4810 Mechatronics Team Project 2 Paul Pounds 8 April 2014 Beginner s Guide to Image Processing or How to make vision-guided killbots step 1 viral infection edition Paul Pounds 8 April 2014 Universit
More informationComputer Vision Lecture 20
Computer Vision Lecture 2 Motion and Optical Flow Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de 28.1.216 Man slides adapted from K. Grauman, S. Seitz, R. Szeliski,
More informationCS 2770: Intro to Computer Vision. Multiple Views. Prof. Adriana Kovashka University of Pittsburgh March 14, 2017
CS 277: Intro to Computer Vision Multiple Views Prof. Adriana Kovashka Universit of Pittsburgh March 4, 27 Plan for toda Affine and projective image transformations Homographies and image mosaics Stereo
More informationMidterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability
Midterm Wed. Local features: detection and description Monday March 7 Prof. UT Austin Covers material up until 3/1 Solutions to practice eam handed out today Bring a 8.5 11 sheet of notes if you want Review
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 12 130228 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Panoramas, Mosaics, Stitching Two View Geometry
More informationWikipedia - Mysid
Wikipedia - Mysid Erik Brynjolfsson, MIT Filtering Edges Corners Feature points Also called interest points, key points, etc. Often described as local features. Szeliski 4.1 Slides from Rick Szeliski,
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationCS4670: Computer Vision
CS4670: Computer Vision Noah Snavely Lecture 6: Feature matching and alignment Szeliski: Chapter 6.1 Reading Last time: Corners and blobs Scale-space blob detector: Example Feature descriptors We know
More informationStereo and Epipolar geometry
Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka
More informationcalibrated coordinates Linear transformation pixel coordinates
1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial
More informationIndex. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263
Index 3D reconstruction, 125 5+1-point algorithm, 284 5-point algorithm, 270 7-point algorithm, 265 8-point algorithm, 263 affine point, 45 affine transformation, 57 affine transformation group, 57 affine
More informationHarder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford
Image matching Harder case by Diva Sian by Diva Sian by scgbt by swashford Even harder case Harder still? How the Afghan Girl was Identified by Her Iris Patterns Read the story NASA Mars Rover images Answer
More informationHarder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford
Image matching Harder case by Diva Sian by Diva Sian by scgbt by swashford Even harder case Harder still? How the Afghan Girl was Identified by Her Iris Patterns Read the story NASA Mars Rover images Answer
More informationStructure from motion
Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t R 2 3,t 3 Camera 1 Camera
More informationCamera Geometry II. COS 429 Princeton University
Camera Geometry II COS 429 Princeton University Outline Projective geometry Vanishing points Application: camera calibration Application: single-view metrology Epipolar geometry Application: stereo correspondence
More informationComputing F class 13. Multiple View Geometry. Comp Marc Pollefeys
Computing F class 3 Multiple View Geometr Comp 90-089 Marc Pollefes Multiple View Geometr course schedule (subject to change) Jan. 7, 9 Intro & motivation Projective D Geometr Jan. 4, 6 (no class) Projective
More informationFinal Exam Study Guide
Final Exam Study Guide Exam Window: 28th April, 12:00am EST to 30th April, 11:59pm EST Description As indicated in class the goal of the exam is to encourage you to review the material from the course.
More informationImage matching. Announcements. Harder case. Even harder case. Project 1 Out today Help session at the end of class. by Diva Sian.
Announcements Project 1 Out today Help session at the end of class Image matching by Diva Sian by swashford Harder case Even harder case How the Afghan Girl was Identified by Her Iris Patterns Read the
More informationEECS 556 Image Processing W 09
EECS 556 Image Processing W 09 Motion estimation Global vs. Local Motion Block Motion Estimation Optical Flow Estimation (normal equation) Man slides of this lecture are courtes of prof Milanfar (UCSC)
More informationPhoto by Carl Warner
Photo b Carl Warner Photo b Carl Warner Photo b Carl Warner Fitting and Alignment Szeliski 6. Computer Vision CS 43, Brown James Has Acknowledgment: Man slides from Derek Hoiem and Grauman&Leibe 2008 AAAI
More information3D Photography: Epipolar geometry
3D Photograph: Epipolar geometr Kalin Kolev, Marc Pollefes Spring 203 http://cvg.ethz.ch/teaching/203spring/3dphoto/ Schedule (tentative) Feb 8 Feb 25 Mar 4 Mar Mar 8 Mar 25 Apr Apr 8 Apr 5 Apr 22 Apr
More informationIndex. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253
Index 3D reconstruction, 123 5+1-point algorithm, 274 5-point algorithm, 260 7-point algorithm, 255 8-point algorithm, 253 affine point, 43 affine transformation, 55 affine transformation group, 55 affine
More informationImage Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58
Image Features: Local Descriptors Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 [Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 2/ 58 Local Features Detection: Identify
More informationApplication questions. Theoretical questions
The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list
More informationTransformations Between Two Images. Translation Rotation Rigid Similarity (scaled rotation) Affine Projective Pseudo Perspective Bi linear
Transformations etween Two Images Translation Rotation Rigid Similarit (scaled rotation) ffine Projective Pseudo Perspective i linear Fundamental Matri Lecture 13 pplications Stereo Structure from Motion
More informationFundamental Matrix. Lecture 13
Fundamental Matri Lecture 13 Transformations etween Two Images Translation Rotation Rigid Similarit (scaled rotation) ffine Projective Pseudo Perspective i linear pplications Stereo Structure from Motion
More informationAutomatic Image Alignment (feature-based)
Automatic Image Alignment (feature-based) Mike Nese with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2006 Today s lecture Feature
More informationBut First: Multi-View Projective Geometry
View Morphing (Seitz & Dyer, SIGGRAPH 96) Virtual Camera Photograph Morphed View View interpolation (ala McMillan) but no depth no camera information Photograph But First: Multi-View Projective Geometry
More informationEpipolar Constraint. Epipolar Lines. Epipolar Geometry. Another look (with math).
Epipolar Constraint Epipolar Lines Potential 3d points Red point - fied => Blue point lies on a line There are 3 degrees of freedom in the position of a point in space; there are four DOF for image points
More informationWarping, Morphing and Mosaics
Computational Photograph and Video: Warping, Morphing and Mosaics Prof. Marc Pollefes Dr. Gabriel Brostow Toda s schedule Last week s recap Warping Morphing Mosaics Toda s schedule Last week s recap Warping
More informationLocal features and image matching. Prof. Xin Yang HUST
Local features and image matching Prof. Xin Yang HUST Last time RANSAC for robust geometric transformation estimation Translation, Affine, Homography Image warping Given a 2D transformation T and a source
More informationCS231A Midterm Review. Friday 5/6/2016
CS231A Midterm Review Friday 5/6/2016 Outline General Logistics Camera Models Non-perspective cameras Calibration Single View Metrology Epipolar Geometry Structure from Motion Active Stereo and Volumetric
More informationBIL Computer Vision Apr 16, 2014
BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm
More informationToday. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography
Computational Photography Matthias Zwicker University of Bern Fall 2009 Today From 2D to 3D using multiple views Introduction Geometry of two views Stereo matching Other applications Multiview geometry
More informationVisual motion. Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys
Visual motion Man slides adapted from S. Seitz, R. Szeliski, M. Pollefes Motion and perceptual organization Sometimes, motion is the onl cue Motion and perceptual organization Sometimes, motion is the
More informationToday s lecture. Image Alignment and Stitching. Readings. Motion models
Today s lecture Image Alignment and Stitching Computer Vision CSE576, Spring 2005 Richard Szeliski Image alignment and stitching motion models cylindrical and spherical warping point-based alignment global
More informationImage Features. Work on project 1. All is Vanity, by C. Allan Gilbert,
Image Features Work on project 1 All is Vanity, by C. Allan Gilbert, 1873-1929 Feature extrac*on: Corners and blobs c Mo*va*on: Automa*c panoramas Credit: Ma9 Brown Why extract features? Mo*va*on: panorama
More informationLocal features: detection and description. Local invariant features
Local features: detection and description Local invariant features Detection of interest points Harris corner detection Scale invariant blob detection: LoG Description of local patches SIFT : Histograms
More informationMAN-522: COMPUTER VISION SET-2 Projections and Camera Calibration
MAN-522: COMPUTER VISION SET-2 Projections and Camera Calibration Image formation How are objects in the world captured in an image? Phsical parameters of image formation Geometric Tpe of projection Camera
More informationObject Recognition with Invariant Features
Object Recognition with Invariant Features Definition: Identify objects or scenes and determine their pose and model parameters Applications Industrial automation and inspection Mobile robots, toys, user
More informationN-Views (1) Homographies and Projection
CS 4495 Computer Vision N-Views (1) Homographies and Projection Aaron Bobick School of Interactive Computing Administrivia PS 2: Get SDD and Normalized Correlation working for a given windows size say
More informationLocal features: detection and description May 12 th, 2015
Local features: detection and description May 12 th, 2015 Yong Jae Lee UC Davis Announcements PS1 grades up on SmartSite PS1 stats: Mean: 83.26 Standard Dev: 28.51 PS2 deadline extended to Saturday, 11:59
More informationStructure from motion
Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t 2 R 3,t 3 Camera 1 Camera
More informationCS231M Mobile Computer Vision Structure from motion
CS231M Mobile Computer Vision Structure from motion - Cameras - Epipolar geometry - Structure from motion Pinhole camera Pinhole perspective projection f o f = focal length o = center of the camera z y
More informationFeature extraction: Corners Harris Corners Pkwy, Charlotte, NC
Featre etraction: Corners 9300 Harris Corners Pkw Charlotte NC Wh etract featres? Motiation: panorama stitching We hae two images how do we combine them? Wh etract featres? Motiation: panorama stitching
More informationComputer Vision Lecture 17
Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 10 130221 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Canny Edge Detector Hough Transform Feature-Based
More informationComputer Vision Lecture 17
Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week
More informationStructure from Motion. Introduction to Computer Vision CSE 152 Lecture 10
Structure from Motion CSE 152 Lecture 10 Announcements Homework 3 is due May 9, 11:59 PM Reading: Chapter 8: Structure from Motion Optional: Multiple View Geometry in Computer Vision, 2nd edition, Hartley
More informationComputational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo --
Computational Optical Imaging - Optique Numerique -- Multiple View Geometry and Stereo -- Winter 2013 Ivo Ihrke with slides by Thorsten Thormaehlen Feature Detection and Matching Wide-Baseline-Matching
More informationFeature descriptors and matching
Feature descriptors and matching Detections at multiple scales Invariance of MOPS Intensity Scale Rotation Color and Lighting Out-of-plane rotation Out-of-plane rotation Better representation than color:
More informationMosaics. Today s Readings
Mosaics VR Seattle: http://www.vrseattle.com/ Full screen panoramas (cubic): http://www.panoramas.dk/ Mars: http://www.panoramas.dk/fullscreen3/f2_mars97.html Today s Readings Szeliski and Shum paper (sections
More informationScale Invariant Feature Transform (SIFT) CS 763 Ajit Rajwade
Scale Invariant Feature Transform (SIFT) CS 763 Ajit Rajwade What is SIFT? It is a technique for detecting salient stable feature points in an image. For ever such point it also provides a set of features
More informationCS 558: Computer Vision 3 rd Set of Notes
1 CS 558: Computer Vision 3 rd Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Overview Denoising Based on slides
More informationAn edge is not a line... Edge Detection. Finding lines in an image. Finding lines in an image. How can we detect lines?
Edge Detection An edge is not a line... original image Cann edge detector Compute image derivatives if gradient magnitude > τ and the value is a local maimum along gradient direction piel is an edge candidate
More informationCamera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration
Camera Calibration Jesus J Caban Schedule! Today:! Camera calibration! Wednesday:! Lecture: Motion & Optical Flow! Monday:! Lecture: Medical Imaging! Final presentations:! Nov 29 th : W. Griffin! Dec 1
More informationEpipolar Geometry and Stereo Vision
Epipolar Geometry and Stereo Vision Computer Vision Jia-Bin Huang, Virginia Tech Many slides from S. Seitz and D. Hoiem Last class: Image Stitching Two images with rotation/zoom but no translation. X x
More informationStep-by-Step Model Buidling
Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure
More informationFeature Detectors and Descriptors: Corners, Lines, etc.
Feature Detectors and Descriptors: Corners, Lines, etc. Edges vs. Corners Edges = maxima in intensity gradient Edges vs. Corners Corners = lots of variation in direction of gradient in a small neighborhood
More informationImage stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac
Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Stitching = alignment + blending geometrical registration
More informationLocal Features: Detection, Description & Matching
Local Features: Detection, Description & Matching Lecture 08 Computer Vision Material Citations Dr George Stockman Professor Emeritus, Michigan State University Dr David Lowe Professor, University of British
More informationUnit 3 Multiple View Geometry
Unit 3 Multiple View Geometry Relations between images of a scene Recovering the cameras Recovering the scene structure http://www.robots.ox.ac.uk/~vgg/hzbook/hzbook1.html 3D structure from images Recover
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem
More informationEpipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz
Epipolar Geometry Prof. D. Stricker With slides from A. Zisserman, S. Lazebnik, Seitz 1 Outline 1. Short introduction: points and lines 2. Two views geometry: Epipolar geometry Relation point/line in two
More informationThere are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...
STEREO VISION The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Bill Freeman and Antonio Torralba (MIT), including their own
More informationMulti-stable Perception. Necker Cube
Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuyuki Kayahara Multiple view geometry Stereo vision Epipolar geometry Lowe Hartley and Zisserman Depth map extraction Essential matrix
More informationMulti-stable Perception. Necker Cube
Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuuki Kaahara Fitting and Alignment Computer Vision Szeliski 6.1 James Has Acknowledgment: Man slides from Derek Hoiem, Lana Lazebnik, and
More informationImage Warping. Many slides from Alyosha Efros + Steve Seitz. Photo by Sean Carroll
Image Warping Man slides from Alosha Efros + Steve Seitz Photo b Sean Carroll Morphing Blend from one object to other with a series of local transformations Image Transformations image filtering: change
More informationEdge and corner detection
Edge and corner detection Prof. Stricker Doz. G. Bleser Computer Vision: Object and People Tracking Goals Where is the information in an image? How is an object characterized? How can I find measurements
More informationCSE 252B: Computer Vision II
CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribes: Jeremy Pollock and Neil Alldrin LECTURE 14 Robust Feature Matching 14.1. Introduction Last lecture we learned how to find interest points
More informationStructure from Motion
11/18/11 Structure from Motion Computer Vision CS 143, Brown James Hays Many slides adapted from Derek Hoiem, Lana Lazebnik, Silvio Saverese, Steve Seitz, and Martial Hebert This class: structure from
More informationCS 532: 3D Computer Vision 7 th Set of Notes
1 CS 532: 3D Computer Vision 7 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Logistics No class on October
More informationStructure from Motion
/8/ Structure from Motion Computer Vision CS 43, Brown James Hays Many slides adapted from Derek Hoiem, Lana Lazebnik, Silvio Saverese, Steve Seitz, and Martial Hebert This class: structure from motion
More informationInstance-level recognition
Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction
More informationHomographies and RANSAC
Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position
More informationLocal invariant features
Local invariant features Tuesday, Oct 28 Kristen Grauman UT-Austin Today Some more Pset 2 results Pset 2 returned, pick up solutions Pset 3 is posted, due 11/11 Local invariant features Detection of interest
More informationComputational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --
Computational Optical Imaging - Optique Numerique -- Single and Multiple View Geometry, Stereo matching -- Autumn 2015 Ivo Ihrke with slides by Thorsten Thormaehlen Reminder: Feature Detection and Matching
More informationLucas-Kanade Image Registration Using Camera Parameters
Lucas-Kanade Image Registration Using Camera Parameters Sunghyun Cho a, Hojin Cho a, Yu-Wing Tai b, Young Su Moon c, Junguk Cho c, Shihwa Lee c, and Seungyong Lee a a POSTECH, Pohang, Korea b KAIST, Daejeon,
More informationCSE 527: Introduction to Computer Vision
CSE 527: Introduction to Computer Vision Week 5 - Class 1: Matching, Stitching, Registration September 26th, 2017 ??? Recap Today Feature Matching Image Alignment Panoramas HW2! Feature Matches Feature
More informationD-Calib: Calibration Software for Multiple Cameras System
D-Calib: Calibration Software for Multiple Cameras Sstem uko Uematsu Tomoaki Teshima Hideo Saito Keio Universit okohama Japan {u-ko tomoaki saito}@ozawa.ics.keio.ac.jp Cao Honghua Librar Inc. Japan cao@librar-inc.co.jp
More informationImage formation - About the course. Grading & Project. Tentative Schedule. Course Content. Students introduction
About the course Instructors: Haibin Ling (hbling@temple, Wachman 305) Hours Lecture: Tuesda 5:30-8:00pm, TTLMAN 403B Office hour: Tuesda 3:00-5:00pm, or b appointment Tetbook Computer Vision: Models,
More information3D Geometry and Camera Calibration
3D Geometr and Camera Calibration 3D Coordinate Sstems Right-handed vs. left-handed 2D Coordinate Sstems ais up vs. ais down Origin at center vs. corner Will often write (u, v) for image coordinates v
More informationToday s lecture. Structure from Motion. Today s lecture. Parameterizing rotations
Today s lecture Structure from Motion Computer Vision CSE576, Spring 2008 Richard Szeliski Geometric camera calibration camera matrix (Direct Linear Transform) non-linear least squares separating intrinsics
More informationCOMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION
COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 7: Image Alignment and Panoramas What s inside your fridge? http://www.cs.washington.edu/education/courses/cse590ss/01wi/ Projection matrix intrinsics projection
More informationFeature Based Registration - Image Alignment
Feature Based Registration - Image Alignment Image Registration Image registration is the process of estimating an optimal transformation between two or more images. Many slides from Alexei Efros http://graphics.cs.cmu.edu/courses/15-463/2007_fall/463.html
More informationVisual Tracking (1) Tracking of Feature Points and Planar Rigid Objects
Intelligent Control Systems Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects Shingo Kagami Graduate School of Information Sciences, Tohoku University swk(at)ic.is.tohoku.ac.jp http://www.ic.is.tohoku.ac.jp/ja/swk/
More informationAutomatic Image Alignment
Automatic Image Alignment with a lot of slides stolen from Steve Seitz and Rick Szeliski Mike Nese CS194: Image Manipulation & Computational Photography Alexei Efros, UC Berkeley, Fall 2018 Live Homography
More informationHartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne
Hartley - Zisserman reading club Part I: Hartley and Zisserman Appendix 6: Iterative estimation methods Part II: Zhengyou Zhang: A Flexible New Technique for Camera Calibration Presented by Daniel Fontijne
More informationInstance-level recognition
Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction
More informationCS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka
CS223b Midterm Exam, Computer Vision Monday February 25th, Winter 2008, Prof. Jana Kosecka Your name email This exam is 8 pages long including cover page. Make sure your exam is not missing any pages.
More informationGeometric camera models and calibration
Geometric camera models and calibration http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 13 Course announcements Homework 3 is out. - Due October
More informationarxiv: v1 [cs.cv] 28 Sep 2018
Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,
More informationStereo Vision. MAN-522 Computer Vision
Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in
More informationRectification and Distortion Correction
Rectification and Distortion Correction Hagen Spies March 12, 2003 Computer Vision Laboratory Department of Electrical Engineering Linköping University, Sweden Contents Distortion Correction Rectification
More informationGeometric Model of Camera
Geometric Model of Camera Dr. Gerhard Roth COMP 42A Winter 25 Version 2 Similar Triangles 2 Geometric Model of Camera Perspective projection P(X,Y,Z) p(,) f X Z f Y Z 3 Parallel lines aren t 4 Figure b
More informationFinal Exam Study Guide CSE/EE 486 Fall 2007
Final Exam Study Guide CSE/EE 486 Fall 2007 Lecture 2 Intensity Sufaces and Gradients Image visualized as surface. Terrain concepts. Gradient of functions in 1D and 2D Numerical derivatives. Taylor series.
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction
More informationMotion Estimation and Optical Flow Tracking
Image Matching Image Retrieval Object Recognition Motion Estimation and Optical Flow Tracking Example: Mosiacing (Panorama) M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 Example 3D Reconstruction
More informationInstance-level recognition part 2
Visual Recognition and Machine Learning Summer School Paris 2011 Instance-level recognition part 2 Josef Sivic http://www.di.ens.fr/~josef INRIA, WILLOW, ENS/INRIA/CNRS UMR 8548 Laboratoire d Informatique,
More information