360 video stabilization
|
|
- Quentin Campbell
- 5 years ago
- Views:
Transcription
1 360 video stabilization PÉTER BODNÁR DEPARTMENT OF IMAGE PROCESSING AND COMPUTER GRAPHICS UNIVERSITY OF SZEGED, HUNGARY
2 Outline About 360 video technology Stabilization: motivation and overview The video stabilization process 2D/3D/Hybrid models Keyframe selection Tracking Rotation estimation Warping Implementation, evaluation Limitations, outlook
3 About 360 degree videos Also known as spherical video, immersive video Recorded with one (omnidirectional) or more cameras (camera rig) Viewer can control the viewing direction during playback Good for action sports (biking, skiing), virtual tours, adult videos
4 Demo (Project Luna)
5 Creation Device options: One camera with multiple lenses More cameras with overlapping angles Most commonly in 4K resolution Output: one spherical video (camera feeds stitched and synchronized)
6 Devices GoPro: since 2002, action cameras, video editing software Omni: rig of 6 HERO cameras ( 5 400) Fusion: 5.2K fully spherical videos, release: september Nokia OZO 2015., (professional fimmakers) 8 lenses, FOV: 195 degrees Samsung Gear 2016., $ fisheye lenses Facebook Surround wide angle lens, 3 fisheye, global shutter N/A Others: LUNA, Ricoh Theta
7 Stitching Multiple (overlapping) images (video frames) panorama Done by camera or software: VideoStitch Kolor: Autopano SkyStitch Steps: Feature detection (corners, blobs, Difference of Gaussian of Harris corners) Registration of images Compositing (single shot, surface) Blending (color/exposure correction, etc)
8
9 Playback Monoscopic (2D) video Direction typically controlled with mouse/touch Stereoscopic (3D) video Google Cardboard (smartphone holder) Oculus
10 Publishing Platforms Youtube (Google), March Oculus (Facebook), September Video making Affordable devices (Samsung Gear 360) Availability of stitching software Video editing made easy Platforms for sharing
11 Comparison: 3D, 360, VR 3D video Depth of field Limited viewpoint 360 video No depth of field Control of viewing direction VR: Mobility Interaction
12 VR sickness flight simulator sickness General discomfort, headache, nausea, vomiting, disorientation Causes Video shaking (hand-held cameras) Too much time spent in VR Personal health conditions Solution Reduce camera shaking more comfortable to watch Make videos smoother in time
13 Stabilization Estimate camera position (x,y,z) during the video Calculate rotation for each video frame Recognize undesired rotations (e.g. shaking) Calculate an inverse model to remove undesired rotations Produce the stabilized video
14 2D video stabilization Feature detection in 2D, narrow Field of View (FOV) Feature matching, low-dimensional similarity, homography transformation estimation Inverse transformation
15 3D stabilization Feature detection on image or in 3D space Estimation of 3D transformation (transformation matrix or relaxed model) Inverse transformation
16 2D: solutions Homography transformation: 1998, Morimoto and Chelappa Cannot handle parallax Mesh warps: 2011, Liu et al. Smoothed flow fields: 2014, Liu et al. Not easy to adapt to 3D spherical video
17 3D: solutions 3D model trajectory and scene estimation Buehler et al, Bhat et al, Liu et al, Kopf et al, Drawbacks Complex process, slow Too specific to some situations
18 Facebook: Hybrid model 3D-2D algorithm Uses deformed rotational model Handles parallax 3D can distinguish between translation and rotation movement 2D fast execution Preserves desired rotations in scene 10-20% file size reduction
19 Hybrid algorithm: overview Feature tracking Keyframe selection Keyframe rotation estimation Keyframe rotation correction Inner frame optimization Relaxation of the model Warping
20 Video frame representation for tracking
21
22 Feature tracking Implemented in cube faces Luminance channel (HSV or YCbCr/Yuv) Tracking by Lukas-Kanade method Points that leave a cube face are dropped Tracked points are converted from 2D locations to 3D unit vectors Further steps are processed on those vectors
23 Lukas-Kanade method Optical flow estimation Sparse features Basic equations for selected pixels and neighborhood Least squares criterion Implemented in OpenCV, MATLAB
24 OpenCV library CV: Computer Vision Open-source Up-to-date Available for many platforms (C++, Java, Python, Android) Modular (Core, GUI, Image Processing, Feature extraction, Machine learning, video) Supports CUDA, OpenGL Well-documented (
25 Shi-Tomasi corner detection Moravec corner detection Harris corner detection Shi-Tomasi (Kanade-Tomasi) Image I, patch (u,v) shifted by (x,y), weighted sum of squared differences S: Taylor expansion of I(u+x, v+y) using partial derivatives: Matrix form:
26 Shi-Tomasi corner detection Eigenvalues show presence of corners as:
27 Keyframe selection Detect features (corners) for each frame First and last frame of the video are keyframes For each frame: Detect features Match features for previous keyframe Count matched points If matched points are less than 50%, mark frame as new keyframe If some time is passed, mark frame as keyframe
28 Rotation estimation Obtain the previously matched point pairs for consecutive key frames Estimate the rotation using Nister s five-point correspondance Use RANSAC procedure Implemented in OpenCV and OpenGV libraries
29 About rotation matrices A matrix that rotates a point in Eucledean space Express rotation around individual axes Consecutive rotations matrix multiplication Rotation between keyframes = rotations for each frame multiplied consecutively Correction of keyframe rotation = inverse of the rotation matrix
30 RANSAC Random Sample Consensus Iterative method for mathematical model creation Inliers: a set of data that makes a model Outliers: data not fitting on the model
31 RANSAC - Algorithm Select a subset of the original data (hypothetical inliers) Fit the model Evaluate the fitted model on all data, create the set of inliers (consensus set) Check the size of the consensus set Repeat fixed number of times Re-estimate the model with best fit
32 RANSAC Advantages and disadvantages Advantages Robust estimation Works with contaminated sets (inliers less than 50%) Has solution for disadvantages (optimal RANSAC) Disadvantages No upper bound on time (exhaustive search) May not produce optimal result
33 Point correspondance, Nister Camera calibration Finding the Essential matrix between corresponding points in images Defines a constraint, assumes pinhole camera model Can be used for pose recovery of calibrated cameras Included in OpenCV Pose recovery Essential matrix decomposition Multiple solutions, cheirality check (positive depth of the triangulated points) Included in OpenCV
34 The OpenGV library Geometric Vision C++ library for calibrated camera pose computation Includes many algorithms + benchmark MATLAB, Python support
35 Inner frame trajectories Linear interpolation between keyframes good initial solution Optimization of the length and smoothness of features
36 The Ceres library Non-linear optimalization and modeling Open-source, C++ Users: Google Microsoft Blender
37 Deformed rotation model
38 About barycentric coordinates Introduced in Also known as areal coordinates Positive within the convex hull X 1 X n : vertices of a simplex in affine space (a 1 + +a n )p = a 1 x a n x n a a n = 1
39 Deformed rotation model
40 Summary
41 Advantages Accuracy: 3D analysis estimates keyframe rotation, separates rotation and translation Robustness: 3D estimation is reliable between keyframes Regularization: Keyframes provide the backbone Deformed rotation model gives smoother result Free from artifacts Speed: Rapid convergence of the 2D inner frame optimization Faster than normal playback
42 Image warping 360 video no cropping Equirectangular equirectangular Rotation estimation: the new position of the pixel Inverting the model: get the value to each position (x,y) from (u,v)
43 Rotation filtering Rotation estimation and removal camera orientation does not change Not all rotations need removal
44 Demo (GoPro Hero5 camera)
45 Two-phase tracking Some patches have high distortion on shaky videos Some features are leaving a specific cube face from the cube map Proposal: repeat tracking on stabilized video further boost of smoothness
46 Hyperlapse Increased speed of videos (typically 3-5x) Rotation estimation and correction is highly recommended Simplest speedup solution: frame dropping More sophisticated solution: Estimate camera velocity Make it smooth: estimate motion vector magnitudes, median + low pass filter, timestamp remap for video frames
47 Demo (Microsoft Hyperlapse)
48 Effect on bitrate Original video can be recovered from modified 360 viewers also apply transformation no additional computation required Bitrate is reduced less rotation between encoded frames
49 Quantitative evaluation (Facebook)
50 Limitations High frequency rolling shutter Wobbling due to the deformed model Shots are not handled, unpredictable Static overlays, logos Person filming the scene
51 Space for improvement Sparse tracking dense tracking Keyframe selection Other methods for rotation estimation and filtering Other representation for frames Using color information
52 References Facebook video stabilization: Devices: Image processing and optimalization libraries: Johannes Kopf: 360 video stabilization:
53 Review 1. Which devices are used for 360 video recording? 2. What is video stitching and how it is performed? 3. What is the difference between 3D video, 360 video and VR? 4. What is video stabilization and what are the steps? 5. What are the main features of OpenCV?
CSE 527: Introduction to Computer Vision
CSE 527: Introduction to Computer Vision Week 5 - Class 1: Matching, Stitching, Registration September 26th, 2017 ??? Recap Today Feature Matching Image Alignment Panoramas HW2! Feature Matches Feature
More informationToday s lecture. Image Alignment and Stitching. Readings. Motion models
Today s lecture Image Alignment and Stitching Computer Vision CSE576, Spring 2005 Richard Szeliski Image alignment and stitching motion models cylindrical and spherical warping point-based alignment global
More informationCSE 252B: Computer Vision II
CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribes: Jeremy Pollock and Neil Alldrin LECTURE 14 Robust Feature Matching 14.1. Introduction Last lecture we learned how to find interest points
More informationcalibrated coordinates Linear transformation pixel coordinates
1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial
More informationMosaics. Today s Readings
Mosaics VR Seattle: http://www.vrseattle.com/ Full screen panoramas (cubic): http://www.panoramas.dk/ Mars: http://www.panoramas.dk/fullscreen3/f2_mars97.html Today s Readings Szeliski and Shum paper (sections
More informationApplication questions. Theoretical questions
The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list
More informationMotion Tracking and Event Understanding in Video Sequences
Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!
More informationCS 4495 Computer Vision Motion and Optic Flow
CS 4495 Computer Vision Aaron Bobick School of Interactive Computing Administrivia PS4 is out, due Sunday Oct 27 th. All relevant lectures posted Details about Problem Set: You may *not* use built in Harris
More informationCOMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE
COMPUTER VISION 2017-2018 > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE OUTLINE Optical flow Lucas-Kanade Horn-Schunck Applications of optical flow Optical flow tracking Histograms of oriented flow Assignment
More informationFeature Tracking and Optical Flow
Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who 1 in turn adapted slides from Steve Seitz, Rick Szeliski,
More informationImage stitching. Announcements. Outline. Image stitching
Announcements Image stitching Project #1 was due yesterday. Project #2 handout will be available on the web later tomorrow. I will set up a webpage for artifact voting soon. Digital Visual Effects, Spring
More informationFeature Tracking and Optical Flow
Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve Seitz, Rick Szeliski,
More informationMatching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.
Matching Compare region of image to region of image. We talked about this for stereo. Important for motion. Epipolar constraint unknown. But motion small. Recognition Find object in image. Recognize object.
More informationIndex. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263
Index 3D reconstruction, 125 5+1-point algorithm, 284 5-point algorithm, 270 7-point algorithm, 265 8-point algorithm, 263 affine point, 45 affine transformation, 57 affine transformation group, 57 affine
More informationIndex. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253
Index 3D reconstruction, 123 5+1-point algorithm, 274 5-point algorithm, 260 7-point algorithm, 255 8-point algorithm, 253 affine point, 43 affine transformation, 55 affine transformation group, 55 affine
More informationINTRODUCTION TO 360 VIDEO. Oliver Wang Adobe Research
INTRODUCTION TO 360 VIDEO Oliver Wang Adobe Research OUTLINE What is 360 video? OUTLINE What is 360 video? How do we represent it? Formats OUTLINE What is 360 video? How do we represent it? How do we create
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationStereo and Epipolar geometry
Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka
More informationMidterm Examination CS 534: Computational Photography
Midterm Examination CS 534: Computational Photography November 3, 2016 NAME: Problem Score Max Score 1 6 2 8 3 9 4 12 5 4 6 13 7 7 8 6 9 9 10 6 11 14 12 6 Total 100 1 of 8 1. [6] (a) [3] What camera setting(s)
More informationAnnouncements. Mosaics. How to do it? Image Mosaics
Announcements Mosaics Project artifact voting Project 2 out today (help session at end of class) http://www.destination36.com/start.htm http://www.vrseattle.com/html/vrview.php?cat_id=&vrs_id=vrs38 Today
More informationAugmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit
Augmented Reality VU Computer Vision 3D Registration (2) Prof. Vincent Lepetit Feature Point-Based 3D Tracking Feature Points for 3D Tracking Much less ambiguous than edges; Point-to-point reprojection
More informationStitching and Blending
Stitching and Blending Kari Pulli VP Computational Imaging Light First project Build your own (basic) programs panorama HDR (really, exposure fusion) The key components register images so their features
More informationIncremental Real-time Bundle Adjustment for Multi-camera Systems with Points at Infinity
Incremental Real-time Bundle Adjustment for Multi-camera Systems with Points at Infinity Johannes Schneider, Thomas Läbe, Wolfgang Förstner 1 Department of Photogrammetry Institute of Geodesy and Geoinformation
More informationImage stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac
Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Stitching = alignment + blending geometrical registration
More informationFinal Exam Study Guide CSE/EE 486 Fall 2007
Final Exam Study Guide CSE/EE 486 Fall 2007 Lecture 2 Intensity Sufaces and Gradients Image visualized as surface. Terrain concepts. Gradient of functions in 1D and 2D Numerical derivatives. Taylor series.
More informationVisual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The
More informationMulti-stable Perception. Necker Cube
Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuyuki Kayahara Multiple view geometry Stereo vision Epipolar geometry Lowe Hartley and Zisserman Depth map extraction Essential matrix
More informationAgenda. Rotations. Camera calibration. Homography. Ransac
Agenda Rotations Camera calibration Homography Ransac Geometric Transformations y x Transformation Matrix # DoF Preserves Icon translation rigid (Euclidean) similarity affine projective h I t h R t h sr
More informationSkybox. Ruoqi He & Chia-Man Hung. February 26, 2016
Skybox Ruoqi He & Chia-Man Hung February 26, 206 Introduction In this project, we present a method to construct a skybox from a series of photos we took ourselves. It is a graphical procedure of creating
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 7: Image Alignment and Panoramas What s inside your fridge? http://www.cs.washington.edu/education/courses/cse590ss/01wi/ Projection matrix intrinsics projection
More informationBSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy
BSB663 Image Processing Pinar Duygulu Slides are adapted from Selim Aksoy Image matching Image matching is a fundamental aspect of many problems in computer vision. Object or scene recognition Solving
More informationCamera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration
Camera Calibration Jesus J Caban Schedule! Today:! Camera calibration! Wednesday:! Lecture: Motion & Optical Flow! Monday:! Lecture: Medical Imaging! Final presentations:! Nov 29 th : W. Griffin! Dec 1
More informationTowards a visual perception system for LNG pipe inspection
Towards a visual perception system for LNG pipe inspection LPV Project Team: Brett Browning (PI), Peter Rander (co PI), Peter Hansen Hatem Alismail, Mohamed Mustafa, Joey Gannon Qri8 Lab A Brief Overview
More informationVisual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania.
Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania 1 What is visual tracking? estimation of the target location over time 2 applications Six main areas:
More informationVisual Tracking (1) Tracking of Feature Points and Planar Rigid Objects
Intelligent Control Systems Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects Shingo Kagami Graduate School of Information Sciences, Tohoku University swk(at)ic.is.tohoku.ac.jp http://www.ic.is.tohoku.ac.jp/ja/swk/
More informationVideo Stabilization with a Depth Camera
Video Stabilization with a Depth Camera Shuaicheng Liu 1 Yinting Wang 1 Lu Yuan 2 Ping Tan 1 Jian Sun 2 1 National University of Singapore 2 Microsoft Research Asia Abstract Previous video stabilization
More informationLecture 16: Computer Vision
CS4442/9542b: Artificial Intelligence II Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field
More informationLecture 16: Computer Vision
CS442/542b: Artificial ntelligence Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field Methods
More informationCS201: Computer Vision Introduction to Tracking
CS201: Computer Vision Introduction to Tracking John Magee 18 November 2014 Slides courtesy of: Diane H. Theriault Question of the Day How can we represent and use motion in images? 1 What is Motion? Change
More informationMulti-View Stereo for Static and Dynamic Scenes
Multi-View Stereo for Static and Dynamic Scenes Wolfgang Burgard Jan 6, 2010 Main references Yasutaka Furukawa and Jean Ponce, Accurate, Dense and Robust Multi-View Stereopsis, 2007 C.L. Zitnick, S.B.
More informationImage processing and features
Image processing and features Gabriele Bleser gabriele.bleser@dfki.de Thanks to Harald Wuest, Folker Wientapper and Marc Pollefeys Introduction Previous lectures: geometry Pose estimation Epipolar geometry
More informationHomographies and RANSAC
Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position
More informationObject Recognition with Invariant Features
Object Recognition with Invariant Features Definition: Identify objects or scenes and determine their pose and model parameters Applications Industrial automation and inspection Mobile robots, toys, user
More informationStep-by-Step Model Buidling
Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure
More informationtechnique: seam carving Image and Video Processing Chapter 9
Chapter 9 Seam Carving for Images and Videos Distributed Algorithms for 2 Introduction Goals Enhance the visual content of images Adapted images should look natural Most relevant content should be clearly
More informationMotion and Optical Flow. Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi
Motion and Optical Flow Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives Motion
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction
More informationIntroduction to Computer Vision
Introduction to Computer Vision Michael J. Black Oct 2009 Motion estimation Goals Motion estimation Affine flow Optimization Large motions Why affine? Monday dense, smooth motion and regularization. Robust
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem
More informationImage Stitching. Slides from Rick Szeliski, Steve Seitz, Derek Hoiem, Ira Kemelmacher, Ali Farhadi
Image Stitching Slides from Rick Szeliski, Steve Seitz, Derek Hoiem, Ira Kemelmacher, Ali Farhadi Combine two or more overlapping images to make one larger image Add example Slide credit: Vaibhav Vaish
More informationIEEE Consumer Electronics Society Calibrating a VR Camera. Adam Rowell CTO, Lucid VR
IEEE Consumer Electronics Society Calibrating a VR Camera Adam Rowell CTO, Lucid VR adam@lucidcam.com Virtual Reality Cameras Lucid VR Camera How Does it Work? Lucid Software Technology Recording: Synchronization
More informationCapturing, Modeling, Rendering 3D Structures
Computer Vision Approach Capturing, Modeling, Rendering 3D Structures Calculate pixel correspondences and extract geometry Not robust Difficult to acquire illumination effects, e.g. specular highlights
More informationRuch (Motion) Rozpoznawanie Obrazów Krzysztof Krawiec Instytut Informatyki, Politechnika Poznańska. Krzysztof Krawiec IDSS
Ruch (Motion) Rozpoznawanie Obrazów Krzysztof Krawiec Instytut Informatyki, Politechnika Poznańska 1 Krzysztof Krawiec IDSS 2 The importance of visual motion Adds entirely new (temporal) dimension to visual
More informationVIDEO STABILIZATION WITH L1-L2 OPTIMIZATION. Hui Qu, Li Song
VIDEO STABILIZATION WITH L-L2 OPTIMIZATION Hui Qu, Li Song Institute of Image Communication and Network Engineering, Shanghai Jiao Tong University ABSTRACT Digital videos often suffer from undesirable
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 12 130228 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Panoramas, Mosaics, Stitching Two View Geometry
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 10 130221 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Canny Edge Detector Hough Transform Feature-Based
More informationCamera Models and Image Formation. Srikumar Ramalingam School of Computing University of Utah
Camera Models and Image Formation Srikumar Ramalingam School of Computing University of Utah srikumar@cs.utah.edu VisualFunHouse.com 3D Street Art Image courtesy: Julian Beaver (VisualFunHouse.com) 3D
More informationCS 378: Autonomous Intelligent Robotics. Instructor: Jivko Sinapov
CS 378: Autonomous Intelligent Robotics Instructor: Jivko Sinapov http://www.cs.utexas.edu/~jsinapov/teaching/cs378/ Visual Registration and Recognition Announcements Homework 6 is out, due 4/5 4/7 Installing
More informationarxiv: v1 [cs.cv] 28 Sep 2018
Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,
More informationLocal Feature Detectors
Local Feature Detectors Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Slides adapted from Cordelia Schmid and David Lowe, CVPR 2003 Tutorial, Matthew Brown,
More informationAdvanced Computer Graphics Transformations. Matthias Teschner
Advanced Computer Graphics Transformations Matthias Teschner Motivation Transformations are used To convert between arbitrary spaces, e.g. world space and other spaces, such as object space, camera space
More informationParameter estimation. Christiano Gava Gabriele Bleser
Parameter estimation Christiano Gava Christiano.Gava@dfki.de Gabriele Bleser gabriele.bleser@dfki.de Introduction Previous lectures: P-matrix 2D projective transformations Estimation (direct linear transform)
More informationCS201 Computer Vision Camera Geometry
CS201 Computer Vision Camera Geometry John Magee 25 November, 2014 Slides Courtesy of: Diane H. Theriault (deht@bu.edu) Question of the Day: How can we represent the relationships between cameras and the
More informationPing Tan. Simon Fraser University
Ping Tan Simon Fraser University Photos vs. Videos (live photos) A good photo tells a story Stories are better told in videos Videos in the Mobile Era (mobile & share) More videos are captured by mobile
More informationOutline. Data Association Scenarios. Data Association Scenarios. Data Association Scenarios
Outline Data Association Scenarios Track Filtering and Gating Global Nearest Neighbor (GNN) Review: Linear Assignment Problem Murthy s k-best Assignments Algorithm Probabilistic Data Association (PDAF)
More informationMotion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation
Motion Sohaib A Khan 1 Introduction So far, we have dealing with single images of a static scene taken by a fixed camera. Here we will deal with sequence of images taken at different time intervals. Motion
More informationStabilizing First Person 360 Degree Videos
Stabilizing First Person 6 Degree Videos Chetan Arora IIIT Delhi Vivek Kwatra Google Research Abstract The use of 6 degree cameras, enabling one to record and share full-spherical 6 8 view without any
More informationOverview. Video. Overview 4/7/2008. Optical flow. Why estimate motion? Motion estimation: Optical flow. Motion Magnification Colorization.
Overview Video Optical flow Motion Magnification Colorization Lecture 9 Optical flow Motion Magnification Colorization Overview Optical flow Combination of slides from Rick Szeliski, Steve Seitz, Alyosha
More informationComputer Vision I Name : CSE 252A, Fall 2012 Student ID : David Kriegman Assignment #1. (Due date: 10/23/2012) x P. = z
Computer Vision I Name : CSE 252A, Fall 202 Student ID : David Kriegman E-Mail : Assignment (Due date: 0/23/202). Perspective Projection [2pts] Consider a perspective projection where a point = z y x P
More informationMonocular Visual-Inertial SLAM. Shaojie Shen Assistant Professor, HKUST Director, HKUST-DJI Joint Innovation Laboratory
Monocular Visual-Inertial SLAM Shaojie Shen Assistant Professor, HKUST Director, HKUST-DJI Joint Innovation Laboratory Why Monocular? Minimum structural requirements Widely available sensors Applications:
More informationOverview. Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion
Overview Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion Binary-Space-Partitioned Images 3-D Surface Extraction from Medical
More informationSIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014
SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image
More informationFeature Based Registration - Image Alignment
Feature Based Registration - Image Alignment Image Registration Image registration is the process of estimating an optimal transformation between two or more images. Many slides from Alexei Efros http://graphics.cs.cmu.edu/courses/15-463/2007_fall/463.html
More informationPin Hole Cameras & Warp Functions
Pin Hole Cameras & Warp Functions Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Pinhole Camera. Homogenous Coordinates. Planar Warp Functions. Example of SLAM for AR Taken from:
More informationReconstruction of Images Distorted by Water Waves
Reconstruction of Images Distorted by Water Waves Arturo Donate and Eraldo Ribeiro Computer Vision Group Outline of the talk Introduction Analysis Background Method Experiments Conclusions Future Work
More informationOptical flow and tracking
EECS 442 Computer vision Optical flow and tracking Intro Optical flow and feature tracking Lucas-Kanade algorithm Motion segmentation Segments of this lectures are courtesy of Profs S. Lazebnik S. Seitz,
More informationGoPro Omni FAQ. Can I use any GoPro HERO4 camera with Omni? Do I need special firmware for the cameras I use with Omni?
GoPro Omni FAQ Can I use any GoPro HERO4 camera with Omni? No, you need to use GoPro HERO4 Black cameras. Other GoPro camera models are not supported. Do I need special firmware for the cameras I use with
More informationMore Mosaic Madness. CS194: Image Manipulation & Computational Photography. Steve Seitz and Rick Szeliski. Jeffrey Martin (jeffrey-martin.
More Mosaic Madness Jeffrey Martin (jeffrey-martin.com) CS194: Image Manipulation & Computational Photography with a lot of slides stolen from Alexei Efros, UC Berkeley, Fall 2018 Steve Seitz and Rick
More informationStructure from Motion
11/18/11 Structure from Motion Computer Vision CS 143, Brown James Hays Many slides adapted from Derek Hoiem, Lana Lazebnik, Silvio Saverese, Steve Seitz, and Martial Hebert This class: structure from
More informationReal-Time Scene Reconstruction. Remington Gong Benjamin Harris Iuri Prilepov
Real-Time Scene Reconstruction Remington Gong Benjamin Harris Iuri Prilepov June 10, 2010 Abstract This report discusses the implementation of a real-time system for scene reconstruction. Algorithms for
More informationVisual Tracking (1) Feature Point Tracking and Block Matching
Intelligent Control Systems Visual Tracking (1) Feature Point Tracking and Block Matching Shingo Kagami Graduate School of Information Sciences, Tohoku University swk(at)ic.is.tohoku.ac.jp http://www.ic.is.tohoku.ac.jp/ja/swk/
More informationSemi-Dense Direct SLAM
Computer Vision Group Technical University of Munich Jakob Engel Jakob Engel, Daniel Cremers David Caruso, Thomas Schöps, Lukas von Stumberg, Vladyslav Usenko, Jörg Stückler, Jürgen Sturm Technical University
More informationWHILE the use of egocentric cameras is on the rise,
1 EgoSampling: Wide View Hyperlapse from Single and Multiple Egocentric Videos Tavi Halperin Yair Poleg Chetan Arora Shmuel Peleg arxiv:1604.07741v1 [cs.cv] 26 Apr 2016 Abstract The possibility of sharing
More informationFeature Matching and RANSAC
Feature Matching and RANSAC Recognising Panoramas. [M. Brown and D. Lowe,ICCV 2003] [Brown, Szeliski, Winder, CVPR 2005] with a lot of slides stolen from Steve Seitz, Rick Szeliski, A. Efros Introduction
More informationAgenda. Rotations. Camera models. Camera calibration. Homographies
Agenda Rotations Camera models Camera calibration Homographies D Rotations R Y = Z r r r r r r r r r Y Z Think of as change of basis where ri = r(i,:) are orthonormal basis vectors r rotated coordinate
More informationProf. Feng Liu. Spring /26/2017
Prof. Feng Liu Spring 2017 http://www.cs.pdx.edu/~fliu/courses/cs510/ 04/26/2017 Last Time Re-lighting HDR 2 Today Panorama Overview Feature detection Mid-term project presentation Not real mid-term 6
More informationMotion Estimation. There are three main types (or applications) of motion estimation:
Members: D91922016 朱威達 R93922010 林聖凱 R93922044 謝俊瑋 Motion Estimation There are three main types (or applications) of motion estimation: Parametric motion (image alignment) The main idea of parametric motion
More information3D Geometry and Camera Calibration
3D Geometry and Camera Calibration 3D Coordinate Systems Right-handed vs. left-handed x x y z z y 2D Coordinate Systems 3D Geometry Basics y axis up vs. y axis down Origin at center vs. corner Will often
More informationPanoramic Image Stitching
Mcgill University Panoramic Image Stitching by Kai Wang Pengbo Li A report submitted in fulfillment for the COMP 558 Final project in the Faculty of Computer Science April 2013 Mcgill University Abstract
More informationAutomatic Image Alignment (feature-based)
Automatic Image Alignment (feature-based) Mike Nese with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2006 Today s lecture Feature
More informationPeripheral drift illusion
Peripheral drift illusion Does it work on other animals? Computer Vision Motion and Optical Flow Many slides adapted from J. Hays, S. Seitz, R. Szeliski, M. Pollefeys, K. Grauman and others Video A video
More informationComputer Vision. Exercise 3 Panorama Stitching 09/12/2013. Compute Vision : Exercise 3 Panorama Stitching
Computer Vision Exercise 3 Panorama Stitching 09/12/2013 Compute Vision : Exercise 3 Panorama Stitching The task Compute Vision : Exercise 3 Panorama Stitching 09/12/2013 2 Pipeline Compute Vision : Exercise
More informationSegmentation and Tracking of Partial Planar Templates
Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract
More information3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava
3D Computer Vision Dense 3D Reconstruction II Prof. Didier Stricker Christiano Gava Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
More informationImage warping and stitching
Image warping and stitching Thurs Oct 15 Last time Feature-based alignment 2D transformations Affine fit RANSAC 1 Robust feature-based alignment Extract features Compute putative matches Loop: Hypothesize
More informationComputer Vision I. Dense Stereo Correspondences. Anita Sellent 1/15/16
Computer Vision I Dense Stereo Correspondences Anita Sellent Stereo Two Cameras Overlapping field of view Known transformation between cameras From disparity compute depth [ Bradski, Kaehler: Learning
More informationCHAPTER 1 INTRODUCTION
1 CHAPTER 1 INTRODUCTION Table of Contents Page No. 1 INTRODUCTION 1.1 Problem overview 2 1.2 Research objective 3 1.3 Thesis outline 7 2 1. INTRODUCTION 1.1 PROBLEM OVERVIEW The process of mapping and
More informationVisual Tracking. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania
Visual Tracking Antonino Furnari Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania furnari@dmi.unict.it 11 giugno 2015 What is visual tracking? estimation
More informationJump Stitch Metadata & Depth Maps Version 1.1
Jump Stitch Metadata & Depth Maps Version 1.1 jump-help@google.com Contents 1. Introduction 1 2. Stitch Metadata File Format 2 3. Coverage Near the Poles 4 4. Coordinate Systems 6 5. Camera Model 6 6.
More informationPerception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich.
Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Perception Sensors Vision Uncertainties, Line extraction
More information