The Video to Panoramic Image Converter
|
|
- Cassandra Hodge
- 5 years ago
- Views:
Transcription
1 The Video to Panoramic Image Converter Mfundo Bill Thesis presented in fulfilment of the requirements for the degree of Bachelor of Science Honours Computer Science at the University of the Western Cape Supervisor: Mehrdad Ghaziasgar Co-supervisor: Reg Dodds This version June 12, 2016
2 ii
3 Declaration I, Mfundo Bill, declare that this thesis The Video to Panoramic Image Converter is my own work, that it has not been submitted before for any degree or assessment at any other university, and that all the sources I have used or quoted have been indicated and acknowledged by means of complete references. Signature: Date: Mfundo Bill. iii
4 iv
5 Abstract This document contains user requirements for the Video to 3D panorama system. This is project chosen as part of the Honours Program in Computer Science at University of the Western Cape. These user requirements where established according to Mr Mehrdad Ghaziasgar, the supervisor of this project. The structure of this document was adapted from the one that was authored by R.P.J. Coset, M.C.G. Leijten, T.J.C. Muller and J.C.J. Mens for the SP- INGRID enviroment system at Eindhoven University of Technology. v
6 vi
7 Key words Computer Vision Video Frames Panorama Panoramic Stitching 3D Modelling Video Summarisation vii
8 viii
9 Acknowledgment This thesis is a compilation of the efforts of many people that helped and me through the years. I would first like to thank my supervisor Mr Mehrdad Ghaziasgar encouraging me during my study. Without our weekly meetings, this work would not have been possible. At this time I would like to extend a very special thanks to the post-graduate assistant. Without their help I would certainly not be where I am today. ix
10 x
11 Contents Declaration Abstract Key words Acknowledgment iii v vii ix 1. User Requirement Document Introduction Scope Definition Overview General Description Main goal General capabilities General constraints User characteristics Users Environment description Assumption and dependencies Use Cases Introduction End User Case Use case Use case Requirements Introduction Functional Requirements Non-Functional Requirements Use Case Diagram User Interface Specification Introduction Description User Interface Behavior xi
12 6. High Level Design Introduction Technical Solution Break Down in Subsystems Subsystems descriptions Input Processing Output Low Level Design Introduction Programming language and Libraries Modules Description Input Processing Gaussian smoothing and Differencing SIFT RANSAC Stitching the images Object Oriented Design Bibliography xii
13 Chapter 1 User Requirement Document 1.1 Introduction The purpose of this document is to specify and describes requirements of the Automatic Video to 3D panorama system Scope The software implements a system that takes in a video, processes it and generates a 3D panoramic image. Given a video, the system must hide all the complicated processing happening and produce the required result Definition Panoramic Frame (of a view or picture) with a wide view surrounding the observer; sweeping. One of many still images that compose a video Image Stitching The process of combining multiple 3D photographic images with overlapping fields of view to produce a segmented panorama or high-resolution image. Commonly performed through the use of computer software, most approaches to image stitching require nearly exact overlaps between images and identical exposures to produce seamless results (Mann and Picard., 1994) Three Dimensional 1
14 Overview Chapter 2 gives a general description for the Video to 3D panorama system. it describes the product, the capabilities and the characteristics of the user and the environments it is going to operate in. Chapter 3 describes a set of use cases. Chapter 4 describes specific requirements for the Video to 3D panorama system and goes on into analysing them.
15 Chapter 2 General Description 2.1 Main goal The Video to 3D panorama system aims to devise an efficient, accurate and user friendly system that converts video into a 3D panoramic image. This can be used for applications like video summarizing and 3D modelling. 2.2 General capabilities The Video to 3D panorama system is designed to convert any kind of video compression. The system uses computer vision techniques to convert video to panorama. The user must only supply the video file and set the time frame at which the panorama must be generated. The Video to 3D panorama system acts as a black box that the user feeds with video content and it gives out a 3D panoramic image as it s output. 2.3 General constraints The Video to 3D panoramic system s quality will be constrained by video noise, video blur,the lighting in the video and the general quality of the video itself. 2.4 User characteristics Users General user: The user that supplies the system with the video file. This user s role is to supply the system with the video that is to be converted and set the time limit of how much of the video should be processed. 3
16 4 2.5 Environment description The system is designed to work in any operating system, ultimately it must also work on the cloud. The system will consist of a user interface that the end user must interact with 2.6 Assumption and dependencies It is assumed that the input video is one continuous scene and has no breaks.
17 Chapter 3 Use Cases 3.1 Introduction A Use Case is a piece of functionality in the system. Those pieces will return a value to a user, where a user does not have to be a human, but could also be a computer system. (R.P.J. Coset, 2006) End User Case Use case 1 The user uploads a video Precondition: true Postcondition: The video being processed User: End User The user uploads a video to the system, sets the time interval for conversion and submits the video Use case 2 The system processes the video. Precondition:Use case 1 User: System Post condition:the final image The system receives a video from the user, processes it and produces the resultant panoramic image. 5
18 6
19 Chapter 4 Requirements 4.1 Introduction Below are the General user requirements of the System Functional Requirements The System must convert a video into a 3 dimensional panoramic image and it must do this efficiently and accurately. The system must be automated as much as possible. The only part of the system that will not be automated is the one where the user uploads the video and sets the time limits Non-Functional Requirements The system must have a user interface that is easy to use and hides all the complex processes happening in the background. 4.2 Use Case Diagram 7
20 8
21 Chapter 5 User Interface Specification 5.1 Introduction This chapter specifically describes the user interface of the video to panorama converter. 5.2 Description The V2PIC system will be a very simple one in the eyes of a user. Since it s goal is to convert an video into a panoramic image, the system is as simple as having one button to convert the specified video to a panorama. 5.3 User Interface Behavior The user must first interact with the interface by providing the path to the video that is to be converted. This can be done by either specifying the path in typing it out into a field in the interface or by interacting with the interface using a mouse and point to where the video is situated. The user then presses the convert button which will trigger the processing and require the user to wait for some time, preferably a short period of time. The user can then view the resultant panorama. 9
22 10
23 Chapter 6 High Level Design 6.1 Introduction This chapter explains the architecture of the system, identifying all the subsystems that form the complete system. 6.2 Technical Solution Break Down in Subsystems The system has 3 main subsystem, the input, processing and output subsystem. This systematic break down will attempt to give a non-technical explanation of how the system works per component. 6.3 Subsystems descriptions This section describes each subsystem and explains what the function is per subsystem Input This component is a simple component that accepts input as the path of the video and passes it to the processing component Processing The processing component on it s own, consists of two components. The first component is one that detects if the video contains a panorama and if so, picks the key frames that will be stitched together. The second subsystem is the one that stitches frames picked by the detector subsystem Output This subsystem is where the resultant panoramic image ends up. This system will help the user view the panorama in an easily navigable way. 11
24 12
25 Chapter 7 Low Level Design 7.1 Introduction This chapter describes and explains each module of the system from a developer s point of view. The description will go into technical details about what each subsystem must do in order to accomplish it s task. 7.2 Programming language and Libraries The programming language used for the system is python3 including the libraries OpenCV and numpy which are available for python3. OpenCV (Open source Computer Vision Library) is used for certain components of the system including reading the video and extracting each frame for analysis. Numpy is used for the computations involved in each process as it is a powerful library for performing matrix operations which is highly required by the system. 7.3 Modules Description Input The first module of the system which is the input component is the most trivial one.it accepts the path of the video and validates if the provided path is correct or with the help of a regular expression and considering that the video exists in the specified path. When all is done as required, the video must then be passed to the processing module Processing The processing module is broken down into two subsystems. The first subsystem is the one which is responsible for selecting the required frames for the stitching processes. The second module is where the image stitching algorithms are apply. The picking of these frames will be based on the frame rate 13
26 14 of the video thus far because the video will be taken in a controlled manner. The stitching subsystem uses the Gaussian smoothing, SIFT and RANSAC algorithms Gaussian smoothing and Differencing This algorithm is used for extracting features of interest. Gaussian smoothing works by taking a square matrix of some chosen size (kernel size), that consists of elements that are numbers that are normally distributed. The matrix (Gaussian kernel) is convolved with the image (Provided that the image is represented by a matrix of 8 bit pixel values) to produce a new image that consists of a smoothed texture with reduced amount of noise compared to the original one. This technique is applied recursively to each resultant image and the result is stored each time. This is repeated for different scales of the image. Differencing the resultant images and finding the extrema gives key-points for features of interest SIFT SIFT, which stands for Scale Invariant Feature Transform helps to describe features from the key-points obtained from the Gaussian smoothing.it does this by selecting some fixed area of pixels surrounding the key point. The output of this algorithm is a four tuple (p, s, r, f) where p is the pixel co-ordinate,
27 15 s is the scale, r is the orientation/direction and f is the feature descriptor. (Ahi, 2016)Below are the steps for the SIFT algorithm: 1. Compute image gradients in the local 16x16 area at a selected scale. 2. Create an array of orientation/direction histograms using the gradients orientations x 4x4 histogram array = 128 dimensions of our feature descriptor. Below is a figure that illustrates orientation calculation for the descriptor: Figure 7.1: Gradient distribution around key-point. The feature descriptor f is view point independant hence it is the same across multiple images if the key-points indicate the same object in the image RANSAC The RANSAC algorithm makes sure that the feature that are the ones that belong in the transformation from image A to image B. In this way RANSAC rolls out all the outliers to make sure that we have the correct features for an overlap of two images. It does this by creating a model (line) for fitting the maximum amount of inliers from the given data points.a threshold for inliers is set and the distant from each point is computed using squared distances from the line to each point. If a point false outside the threshold, then the point is labeled as an outlier. (Fiscler, 1905) Below are the steps for the RANSAC algorithm: 1. Randomly Sample the number of data points required to fit the model
28 16 2. Comp-ute model parameters using the sample. 3. Score the fraction of inliers within a present threshold of the model. 4. Repeat until model is found. Figure 7.2: RANSAC with inliers and outliers illustration Stitching the images Using the RANSAC algorithm, given Image A and Image B, extracted features from A are compared with the ones from B. The comparison of these features determines which ones overlap (are inliers) and which ones do not (outliers). The features that overlap belong in our transformation model hypothesis. We repeat this process until with the number of repetitions depending on the probability at which we have at least one outlier.we then stitching the Images together by matching the key-points (pixel co-ordinate) where the features are situated according to the feature descriptors. The last step is to stitch each selected frame to the each resultant panorama. This is done as an iterative process until all frames are exhausted. Below is a rough overall algorithm for the stitching.
29 17 Data: Video Result: Panaroma prevframe = Video.Cputure; while Video still playing do set frame rate; currframe = frame from video capture; keypointa = GaussianDiff(prevFrame); keypointb = GaussianDiff(currFrame); descriptora = SIFT(keyPointA); descriptorb = SIFT(keyPointB); matchedset = Matcher(DescriptorA,DescriptorB); resultantimage = Stitcher(matchedSet); prevframe = resultantimage; end Algorithm 1: Algorithm for the stitching process Figure 7.3: Key-point Matching illustration.
30 Object Oriented Design Below is thus far the object oriented structure of how all the modules of the system will work. Figure 7.4: Object Oriented Design
31 Bibliography Ahi, K. (2016). Modeling of terahertz images based on x-ray images: a novel approach for verification of terahertz images and identification of objects with fine details beyond terahertz resolution. Conference Paper, page 3. Fiscler, M. A. (1905). Random Sampling Consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Technical Note, (213):3. Mann, S. and Picard., R. W. (1994). Virtual bellows: constructing highquality images from video., in proceedings of the ieee first international conference on image processingaustin. Technical report. R.P.J. Coset, M.C.G. Leijten, T. M. J. M. (2006). User requirements document,spingrid. Technical report, Technische Informatica, Eindhoven University of Technology. 19
Panoramic Image Stitching
Mcgill University Panoramic Image Stitching by Kai Wang Pengbo Li A report submitted in fulfillment for the COMP 558 Final project in the Faculty of Computer Science April 2013 Mcgill University Abstract
More informationBSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy
BSB663 Image Processing Pinar Duygulu Slides are adapted from Selim Aksoy Image matching Image matching is a fundamental aspect of many problems in computer vision. Object or scene recognition Solving
More informationMidterm Examination CS 534: Computational Photography
Midterm Examination CS 534: Computational Photography November 3, 2016 NAME: Problem Score Max Score 1 6 2 8 3 9 4 12 5 4 6 13 7 7 8 6 9 9 10 6 11 14 12 6 Total 100 1 of 8 1. [6] (a) [3] What camera setting(s)
More informationFeature Matching and RANSAC
Feature Matching and RANSAC Recognising Panoramas. [M. Brown and D. Lowe,ICCV 2003] [Brown, Szeliski, Winder, CVPR 2005] with a lot of slides stolen from Steve Seitz, Rick Szeliski, A. Efros Introduction
More informationLocal Feature Detectors
Local Feature Detectors Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Slides adapted from Cordelia Schmid and David Lowe, CVPR 2003 Tutorial, Matthew Brown,
More informationHomographies and RANSAC
Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position
More informationUncertainties: Representation and Propagation & Line Extraction from Range data
41 Uncertainties: Representation and Propagation & Line Extraction from Range data 42 Uncertainty Representation Section 4.1.3 of the book Sensing in the real world is always uncertain How can uncertainty
More informationFeature Based Registration - Image Alignment
Feature Based Registration - Image Alignment Image Registration Image registration is the process of estimating an optimal transformation between two or more images. Many slides from Alexei Efros http://graphics.cs.cmu.edu/courses/15-463/2007_fall/463.html
More informationComputer Vision. Exercise 3 Panorama Stitching 09/12/2013. Compute Vision : Exercise 3 Panorama Stitching
Computer Vision Exercise 3 Panorama Stitching 09/12/2013 Compute Vision : Exercise 3 Panorama Stitching The task Compute Vision : Exercise 3 Panorama Stitching 09/12/2013 2 Pipeline Compute Vision : Exercise
More informationCSE 527: Introduction to Computer Vision
CSE 527: Introduction to Computer Vision Week 5 - Class 1: Matching, Stitching, Registration September 26th, 2017 ??? Recap Today Feature Matching Image Alignment Panoramas HW2! Feature Matches Feature
More informationSIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014
SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image
More informationChapter 3 Image Registration. Chapter 3 Image Registration
Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation
More informationObservations. Basic iteration Line estimated from 2 inliers
Line estimated from 2 inliers 3 Observations We need (in this case!) a minimum of 2 points to determine a line Given such a line l, we can determine how well any other point y fits the line l For example:
More informationImage Stitching using Watersheds and Graph Cuts
Image Stitching using Watersheds and Graph Cuts Patrik Nyman Centre for Mathematical Sciences, Lund University, Sweden patnym@maths.lth.se 1. Introduction Image stitching is commonly used in many different
More informationImage stitching. Announcements. Outline. Image stitching
Announcements Image stitching Project #1 was due yesterday. Project #2 handout will be available on the web later tomorrow. I will set up a webpage for artifact voting soon. Digital Visual Effects, Spring
More informationFrom Structure-from-Motion Point Clouds to Fast Location Recognition
From Structure-from-Motion Point Clouds to Fast Location Recognition Arnold Irschara1;2, Christopher Zach2, Jan-Michael Frahm2, Horst Bischof1 1Graz University of Technology firschara, bischofg@icg.tugraz.at
More informationMosaics. Today s Readings
Mosaics VR Seattle: http://www.vrseattle.com/ Full screen panoramas (cubic): http://www.panoramas.dk/ Mars: http://www.panoramas.dk/fullscreen3/f2_mars97.html Today s Readings Szeliski and Shum paper (sections
More informationLocation Based Selling Platform for Mobile Buyers
Location Based Selling Platform for Mobile Buyers M. M. Buddhika Mawella 149219M Faculty of Information Technology University of Moratuwa April 2017 Location Based Selling Platform for Mobile Buyers M.
More informationReal Time Multi-Sensor Data Acquisition and Processing for a Road Mapping System
Real Time Multi-Sensor Data Acquisition and Processing for a Road Mapping System by Xiang Luo A thesis submitted for the degree of Master of Engineering (Research) Faculty of Engineering and Information
More informationImage Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58
Image Features: Local Descriptors Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 [Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 2/ 58 Local Features Detection: Identify
More informationAdvanced Video Content Analysis and Video Compression (5LSH0), Module 4
Advanced Video Content Analysis and Video Compression (5LSH0), Module 4 Visual feature extraction Part I: Color and texture analysis Sveta Zinger Video Coding and Architectures Research group, TU/e ( s.zinger@tue.nl
More informationImage correspondences and structure from motion
Image correspondences and structure from motion http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 20 Course announcements Homework 5 posted.
More informationAnalysis of Image and Video Using Color, Texture and Shape Features for Object Identification
IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 16, Issue 6, Ver. VI (Nov Dec. 2014), PP 29-33 Analysis of Image and Video Using Color, Texture and Shape Features
More informationRobotics Programming Laboratory
Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car
More informationA Comparison of SIFT, PCA-SIFT and SURF
A Comparison of SIFT, PCA-SIFT and SURF Luo Juan Computer Graphics Lab, Chonbuk National University, Jeonju 561-756, South Korea qiuhehappy@hotmail.com Oubong Gwun Computer Graphics Lab, Chonbuk National
More informationCS1114: Study Guide 2
CS4: Study Guide 2 This document covers the topics we ve covered in the second part of the course. Please refer to the class slides for more details. Polygons and convex hulls A polygon is a set of 2D
More informationHough Transform and RANSAC
CS4501: Introduction to Computer Vision Hough Transform and RANSAC Various slides from previous courses by: D.A. Forsyth (Berkeley / UIUC), I. Kokkinos (Ecole Centrale / UCL). S. Lazebnik (UNC / UIUC),
More informationPerception IV: Place Recognition, Line Extraction
Perception IV: Place Recognition, Line Extraction Davide Scaramuzza University of Zurich Margarita Chli, Paul Furgale, Marco Hutter, Roland Siegwart 1 Outline of Today s lecture Place recognition using
More informationObject Recognition with Invariant Features
Object Recognition with Invariant Features Definition: Identify objects or scenes and determine their pose and model parameters Applications Industrial automation and inspection Mobile robots, toys, user
More informationImage stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac
Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Stitching = alignment + blending geometrical registration
More informationLocal Features Tutorial: Nov. 8, 04
Local Features Tutorial: Nov. 8, 04 Local Features Tutorial References: Matlab SIFT tutorial (from course webpage) Lowe, David G. Distinctive Image Features from Scale Invariant Features, International
More informationLocal invariant features
Local invariant features Tuesday, Oct 28 Kristen Grauman UT-Austin Today Some more Pset 2 results Pset 2 returned, pick up solutions Pset 3 is posted, due 11/11 Local invariant features Detection of interest
More informationA Summary of Projective Geometry
A Summary of Projective Geometry Copyright 22 Acuity Technologies Inc. In the last years a unified approach to creating D models from multiple images has been developed by Beardsley[],Hartley[4,5,9],Torr[,6]
More informationScale Invariant Feature Transform
Scale Invariant Feature Transform Why do we care about matching features? Camera calibration Stereo Tracking/SFM Image moiaicing Object/activity Recognition Objection representation and recognition Image
More informationDepartment of Computer Science University of Western Cape Computer Science Honours Project: Anaglyph Video By: Jihaad Pienaar Supervisors: Mehrdad
Department of Computer Science University of Western Cape Computer Science Honours Project: Anaglyph Video By: Jihaad Pienaar Supervisors: Mehrdad Ghaziasgar and James Connan Glossary Depth Map Stereo
More informationPanoramic Image Mosaicing
Panoramic Image Mosaicing Mrs. Hetal M.Patel 1, Asst. Prof. Pinal. J. Patel 2, Asst. Prof. Mr. Sandip G. Patel 3 1 C.S.E. Department, Government College of Engineering, Gandhinagar Gujarat Technology University,
More informationStructured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov
Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter
More informationA NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION
A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM Karthik Krish Stuart Heinrich Wesley E. Snyder Halil Cakir Siamak Khorram North Carolina State University Raleigh, 27695 kkrish@ncsu.edu sbheinri@ncsu.edu
More informationMulti-stable Perception. Necker Cube
Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuyuki Kayahara Multiple view geometry Stereo vision Epipolar geometry Lowe Hartley and Zisserman Depth map extraction Essential matrix
More informationObject Reconstruction
B. Scholz Object Reconstruction 1 / 39 MIN-Fakultät Fachbereich Informatik Object Reconstruction Benjamin Scholz Universität Hamburg Fakultät für Mathematik, Informatik und Naturwissenschaften Fachbereich
More information[Programming Assignment] (1)
http://crcv.ucf.edu/people/faculty/bagci/ [Programming Assignment] (1) Computer Vision Dr. Ulas Bagci (Fall) 2015 University of Central Florida (UCF) Coding Standard and General Requirements Code for all
More informationFeature Detectors and Descriptors: Corners, Lines, etc.
Feature Detectors and Descriptors: Corners, Lines, etc. Edges vs. Corners Edges = maxima in intensity gradient Edges vs. Corners Corners = lots of variation in direction of gradient in a small neighborhood
More informationPerception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich.
Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Perception Sensors Vision Uncertainties, Line extraction
More informationA System of Image Matching and 3D Reconstruction
A System of Image Matching and 3D Reconstruction CS231A Project Report 1. Introduction Xianfeng Rui Given thousands of unordered images of photos with a variety of scenes in your gallery, you will find
More informationGlobal Flow Estimation. Lecture 9
Motion Models Image Transformations to relate two images 3D Rigid motion Perspective & Orthographic Transformation Planar Scene Assumption Transformations Translation Rotation Rigid Affine Homography Pseudo
More informationThe SIFT (Scale Invariant Feature
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor developed by David Lowe University of British Columbia Initial paper ICCV 1999 Newer journal paper IJCV 2004 Review: Matt Brown s Canonical
More informationProf. Noah Snavely CS Administrivia. A4 due on Friday (please sign up for demo slots)
Robust fitting Prof. Noah Snavely CS111 http://www.cs.cornell.edu/courses/cs111 Administrivia A due on Friday (please sign up for demo slots) A5 will be out soon Prelim is coming up, Tuesday, / Roadmap
More informationAutomatic Image Alignment (feature-based)
Automatic Image Alignment (feature-based) Mike Nese with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2006 Today s lecture Feature
More informationURBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES
URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES An Undergraduate Research Scholars Thesis by RUI LIU Submitted to Honors and Undergraduate Research Texas A&M University in partial fulfillment
More informationScale Invariant Feature Transform
Why do we care about matching features? Scale Invariant Feature Transform Camera calibration Stereo Tracking/SFM Image moiaicing Object/activity Recognition Objection representation and recognition Automatic
More informationComputer Vision, Assignment 4 Model Fitting
Centre for Mathematical Sciences, February 2013 Due 2013-02-26 Computer Vision, Assignment 4 Model Fitting 1 Instructions In this assignment you will study model fitting. In particular you will use random
More informationStitching and Blending
Stitching and Blending Kari Pulli VP Computational Imaging Light First project Build your own (basic) programs panorama HDR (really, exposure fusion) The key components register images so their features
More informationCLASSIFICATION AND CHANGE DETECTION
IMAGE ANALYSIS, CLASSIFICATION AND CHANGE DETECTION IN REMOTE SENSING With Algorithms for ENVI/IDL and Python THIRD EDITION Morton J. Canty CRC Press Taylor & Francis Group Boca Raton London NewYork CRC
More informationCSE 252B: Computer Vision II
CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribes: Jeremy Pollock and Neil Alldrin LECTURE 14 Robust Feature Matching 14.1. Introduction Last lecture we learned how to find interest points
More informationComparison of Feature Detection and Matching Approaches: SIFT and SURF
GRD Journals- Global Research and Development Journal for Engineering Volume 2 Issue 4 March 2017 ISSN: 2455-5703 Comparison of Detection and Matching Approaches: SIFT and SURF Darshana Mistry PhD student
More informationCS 558: Computer Vision 4 th Set of Notes
1 CS 558: Computer Vision 4 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Overview Keypoint matching Hessian
More informationDigital Image Processing
Digital Image Processing Third Edition Rafael C. Gonzalez University of Tennessee Richard E. Woods MedData Interactive PEARSON Prentice Hall Pearson Education International Contents Preface xv Acknowledgments
More informationAutomatic Image Alignment
Automatic Image Alignment Mike Nese with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2010 Live Homography DEMO Check out panoramio.com
More informationIntroduction to Medical Imaging (5XSA0)
1 Introduction to Medical Imaging (5XSA0) Visual feature extraction Color and texture analysis Sveta Zinger ( s.zinger@tue.nl ) Introduction (1) Features What are features? Feature a piece of information
More informationGesture based PTZ camera control
Gesture based PTZ camera control Report submitted in May 2014 to the department of Computer Science and Engineering of National Institute of Technology Rourkela in partial fulfillment of the requirements
More informationINFO0948 Fitting and Shape Matching
INFO0948 Fitting and Shape Matching Renaud Detry University of Liège, Belgium Updated March 31, 2015 1 / 33 These slides are based on the following book: D. Forsyth and J. Ponce. Computer vision: a modern
More information3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.
3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction
More informationModel Fitting, RANSAC. Jana Kosecka
Model Fitting, RANSAC Jana Kosecka Fitting: Overview If we know which points belong to the line, how do we find the optimal line parameters? Least squares What if there are outliers? Robust fitting, RANSAC
More information521466S Machine Vision Assignment #3 Image Features
521466S Machine Vision Assignment #3 Image Features Spring 2018 This assignment explores feature detection, extraction, and matching. We will implement the well-known Harris corner detector and use Matlab
More informationImplementing the Scale Invariant Feature Transform(SIFT) Method
Implementing the Scale Invariant Feature Transform(SIFT) Method YU MENG and Dr. Bernard Tiddeman(supervisor) Department of Computer Science University of St. Andrews yumeng@dcs.st-and.ac.uk Abstract The
More informationAn Algorithm for Seamless Image Stitching and Its Application
An Algorithm for Seamless Image Stitching and Its Application Jing Xing, Zhenjiang Miao, and Jing Chen Institute of Information Science, Beijing JiaoTong University, Beijing 100044, P.R. China Abstract.
More informationA Keypoint Descriptor Inspired by Retinal Computation
A Keypoint Descriptor Inspired by Retinal Computation Bongsoo Suh, Sungjoon Choi, Han Lee Stanford University {bssuh,sungjoonchoi,hanlee}@stanford.edu Abstract. The main goal of our project is to implement
More informationInstance-level recognition
Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction
More informationCS 378: Autonomous Intelligent Robotics. Instructor: Jivko Sinapov
CS 378: Autonomous Intelligent Robotics Instructor: Jivko Sinapov http://www.cs.utexas.edu/~jsinapov/teaching/cs378/ Visual Registration and Recognition Announcements Homework 6 is out, due 4/5 4/7 Installing
More informationSchool of Computing University of Utah
School of Computing University of Utah Presentation Outline 1 2 3 4 Main paper to be discussed David G. Lowe, Distinctive Image Features from Scale-Invariant Keypoints, IJCV, 2004. How to find useful keypoints?
More informationImage Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner (tanner@vision.ee.ethz.ch) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
More informationRANSAC and some HOUGH transform
RANSAC and some HOUGH transform Thank you for the slides. They come mostly from the following source Dan Huttenlocher Cornell U Matching and Fitting Recognition and matching are closely related to fitting
More informationA Novel Extreme Point Selection Algorithm in SIFT
A Novel Extreme Point Selection Algorithm in SIFT Ding Zuchun School of Electronic and Communication, South China University of Technolog Guangzhou, China zucding@gmail.com Abstract. This paper proposes
More informationMotivation. Intensity Levels
Motivation Image Intensity and Point Operations Dr. Edmund Lam Department of Electrical and Electronic Engineering The University of Hong ong A digital image is a matrix of numbers, each corresponding
More information3D OBJECT RECONTRUCTION USING MULTIPLE-VIEW GEOMETRY: SIFT DETECTION LEOW RUEY SHYAN
3D OBJECT RECONTRUCTION USING MULTIPLE-VIEW GEOMETRY: SIFT DETECTION LEOW RUEY SHYAN A project report submitted in partial fulfilment of the requirements for the award of the degree of Bachelor (Hons.)
More informationUNIVERSITI MALAYSIA PAHANG
IMAGE ENHANCEMENT AND SEGMENTATION ON SIMULTANEOUS LATENT FINGERPRINT DETECTION ROZITA BINTI MOHD YUSOF MASTER OF COMPUTER SCIENCE UNIVERSITI MALAYSIA PAHANG IMAGE ENHANCEMENT AND SEGMENTATION ON SIMULTANEOUS
More informationCombining Appearance and Topology for Wide
Combining Appearance and Topology for Wide Baseline Matching Dennis Tell and Stefan Carlsson Presented by: Josh Wills Image Point Correspondences Critical foundation for many vision applications 3-D reconstruction,
More informationAccurate Image Registration from Local Phase Information
Accurate Image Registration from Local Phase Information Himanshu Arora, Anoop M. Namboodiri, and C.V. Jawahar Center for Visual Information Technology, IIIT, Hyderabad, India { himanshu@research., anoop@,
More informationMiniature faking. In close-up photo, the depth of field is limited.
Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg
More informationFace and Nose Detection in Digital Images using Local Binary Patterns
Face and Nose Detection in Digital Images using Local Binary Patterns Stanko Kružić Post-graduate student University of Split, Faculty of Electrical Engineering, Mechanical Engineering and Naval Architecture
More informationThe Lucas & Kanade Algorithm
The Lucas & Kanade Algorithm Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Registration, Registration, Registration. Linearizing Registration. Lucas & Kanade Algorithm. 3 Biggest
More informationFundamentals of Digital Image Processing
\L\.6 Gw.i Fundamentals of Digital Image Processing A Practical Approach with Examples in Matlab Chris Solomon School of Physical Sciences, University of Kent, Canterbury, UK Toby Breckon School of Engineering,
More informationImage stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac
Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Stitching = alignment + blending geometrical registration
More information3D Environment Reconstruction
3D Environment Reconstruction Using Modified Color ICP Algorithm by Fusion of a Camera and a 3D Laser Range Finder The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October 11-15,
More informationComputer Vision I - Filtering and Feature detection
Computer Vision I - Filtering and Feature detection Carsten Rother 30/10/2015 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image
More informationSrikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah
School of Computing University of Utah Presentation Outline 1 2 3 Forward Projection (Reminder) u v 1 KR ( I t ) X m Y m Z m 1 Backward Projection (Reminder) Q K 1 q Presentation Outline 1 2 3 Sample Problem
More informationComputer Vision CSCI-GA Assignment 1.
Computer Vision CSCI-GA.2272-001 Assignment 1. September 22, 2017 Introduction This assignment explores various methods for aligning images and feature extraction. There are four parts to the assignment:
More informationSUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS
SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS Cognitive Robotics Original: David G. Lowe, 004 Summary: Coen van Leeuwen, s1460919 Abstract: This article presents a method to extract
More informationFeature Matching and Robust Fitting
Feature Matching and Robust Fitting Computer Vision CS 143, Brown Read Szeliski 4.1 James Hays Acknowledgment: Many slides from Derek Hoiem and Grauman&Leibe 2008 AAAI Tutorial Project 2 questions? This
More informationComputer Vision for HCI. Topics of This Lecture
Computer Vision for HCI Interest Points Topics of This Lecture Local Invariant Features Motivation Requirements, Invariances Keypoint Localization Features from Accelerated Segment Test (FAST) Harris Shi-Tomasi
More informationEVALUATION OF SEQUENTIAL IMAGES FOR PHOTOGRAMMETRICALLY POINT DETERMINATION
Archives of Photogrammetry, Cartography and Remote Sensing, Vol. 22, 2011, pp. 285-296 ISSN 2083-2214 EVALUATION OF SEQUENTIAL IMAGES FOR PHOTOGRAMMETRICALLY POINT DETERMINATION Michał Kowalczyk 1 1 Department
More informationObtaining Feature Correspondences
Obtaining Feature Correspondences Neill Campbell May 9, 2008 A state-of-the-art system for finding objects in images has recently been developed by David Lowe. The algorithm is termed the Scale-Invariant
More informationFeature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking
Feature descriptors Alain Pagani Prof. Didier Stricker Computer Vision: Object and People Tracking 1 Overview Previous lectures: Feature extraction Today: Gradiant/edge Points (Kanade-Tomasi + Harris)
More informationLocal features: detection and description May 12 th, 2015
Local features: detection and description May 12 th, 2015 Yong Jae Lee UC Davis Announcements PS1 grades up on SmartSite PS1 stats: Mean: 83.26 Standard Dev: 28.51 PS2 deadline extended to Saturday, 11:59
More informationImage processing and features
Image processing and features Gabriele Bleser gabriele.bleser@dfki.de Thanks to Harald Wuest, Folker Wientapper and Marc Pollefeys Introduction Previous lectures: geometry Pose estimation Epipolar geometry
More informationAn Implementation on Histogram of Oriented Gradients for Human Detection
An Implementation on Histogram of Oriented Gradients for Human Detection Cansın Yıldız Dept. of Computer Engineering Bilkent University Ankara,Turkey cansin@cs.bilkent.edu.tr Abstract I implemented a Histogram
More informationCITS 4402 Computer Vision
CITS 4402 Computer Vision A/Prof Ajmal Mian Adj/A/Prof Mehdi Ravanbakhsh, CEO at Mapizy (www.mapizy.com) and InFarm (www.infarm.io) Lecture 02 Binary Image Analysis Objectives Revision of image formation
More informationCS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry
CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry Some material taken from: David Lowe, UBC Jiri Matas, CMP Prague http://cmp.felk.cvut.cz/~matas/papers/presentations/matas_beyondransac_cvprac05.ppt
More informationTA Section 7 Problem Set 3. SIFT (Lowe 2004) Shape Context (Belongie et al. 2002) Voxel Coloring (Seitz and Dyer 1999)
TA Section 7 Problem Set 3 SIFT (Lowe 2004) Shape Context (Belongie et al. 2002) Voxel Coloring (Seitz and Dyer 1999) Sam Corbett-Davies TA Section 7 02-13-2014 Distinctive Image Features from Scale-Invariant
More informationClassification of objects from Video Data (Group 30)
Classification of objects from Video Data (Group 30) Sheallika Singh 12665 Vibhuti Mahajan 12792 Aahitagni Mukherjee 12001 M Arvind 12385 1 Motivation Video surveillance has been employed for a long time
More information