The Video to Panoramic Image Converter

Size: px
Start display at page:

Download "The Video to Panoramic Image Converter"

Transcription

1 The Video to Panoramic Image Converter Mfundo Bill Thesis presented in fulfilment of the requirements for the degree of Bachelor of Science Honours Computer Science at the University of the Western Cape Supervisor: Mehrdad Ghaziasgar Co-supervisor: Reg Dodds This version June 12, 2016

2 ii

3 Declaration I, Mfundo Bill, declare that this thesis The Video to Panoramic Image Converter is my own work, that it has not been submitted before for any degree or assessment at any other university, and that all the sources I have used or quoted have been indicated and acknowledged by means of complete references. Signature: Date: Mfundo Bill. iii

4 iv

5 Abstract This document contains user requirements for the Video to 3D panorama system. This is project chosen as part of the Honours Program in Computer Science at University of the Western Cape. These user requirements where established according to Mr Mehrdad Ghaziasgar, the supervisor of this project. The structure of this document was adapted from the one that was authored by R.P.J. Coset, M.C.G. Leijten, T.J.C. Muller and J.C.J. Mens for the SP- INGRID enviroment system at Eindhoven University of Technology. v

6 vi

7 Key words Computer Vision Video Frames Panorama Panoramic Stitching 3D Modelling Video Summarisation vii

8 viii

9 Acknowledgment This thesis is a compilation of the efforts of many people that helped and me through the years. I would first like to thank my supervisor Mr Mehrdad Ghaziasgar encouraging me during my study. Without our weekly meetings, this work would not have been possible. At this time I would like to extend a very special thanks to the post-graduate assistant. Without their help I would certainly not be where I am today. ix

10 x

11 Contents Declaration Abstract Key words Acknowledgment iii v vii ix 1. User Requirement Document Introduction Scope Definition Overview General Description Main goal General capabilities General constraints User characteristics Users Environment description Assumption and dependencies Use Cases Introduction End User Case Use case Use case Requirements Introduction Functional Requirements Non-Functional Requirements Use Case Diagram User Interface Specification Introduction Description User Interface Behavior xi

12 6. High Level Design Introduction Technical Solution Break Down in Subsystems Subsystems descriptions Input Processing Output Low Level Design Introduction Programming language and Libraries Modules Description Input Processing Gaussian smoothing and Differencing SIFT RANSAC Stitching the images Object Oriented Design Bibliography xii

13 Chapter 1 User Requirement Document 1.1 Introduction The purpose of this document is to specify and describes requirements of the Automatic Video to 3D panorama system Scope The software implements a system that takes in a video, processes it and generates a 3D panoramic image. Given a video, the system must hide all the complicated processing happening and produce the required result Definition Panoramic Frame (of a view or picture) with a wide view surrounding the observer; sweeping. One of many still images that compose a video Image Stitching The process of combining multiple 3D photographic images with overlapping fields of view to produce a segmented panorama or high-resolution image. Commonly performed through the use of computer software, most approaches to image stitching require nearly exact overlaps between images and identical exposures to produce seamless results (Mann and Picard., 1994) Three Dimensional 1

14 Overview Chapter 2 gives a general description for the Video to 3D panorama system. it describes the product, the capabilities and the characteristics of the user and the environments it is going to operate in. Chapter 3 describes a set of use cases. Chapter 4 describes specific requirements for the Video to 3D panorama system and goes on into analysing them.

15 Chapter 2 General Description 2.1 Main goal The Video to 3D panorama system aims to devise an efficient, accurate and user friendly system that converts video into a 3D panoramic image. This can be used for applications like video summarizing and 3D modelling. 2.2 General capabilities The Video to 3D panorama system is designed to convert any kind of video compression. The system uses computer vision techniques to convert video to panorama. The user must only supply the video file and set the time frame at which the panorama must be generated. The Video to 3D panorama system acts as a black box that the user feeds with video content and it gives out a 3D panoramic image as it s output. 2.3 General constraints The Video to 3D panoramic system s quality will be constrained by video noise, video blur,the lighting in the video and the general quality of the video itself. 2.4 User characteristics Users General user: The user that supplies the system with the video file. This user s role is to supply the system with the video that is to be converted and set the time limit of how much of the video should be processed. 3

16 4 2.5 Environment description The system is designed to work in any operating system, ultimately it must also work on the cloud. The system will consist of a user interface that the end user must interact with 2.6 Assumption and dependencies It is assumed that the input video is one continuous scene and has no breaks.

17 Chapter 3 Use Cases 3.1 Introduction A Use Case is a piece of functionality in the system. Those pieces will return a value to a user, where a user does not have to be a human, but could also be a computer system. (R.P.J. Coset, 2006) End User Case Use case 1 The user uploads a video Precondition: true Postcondition: The video being processed User: End User The user uploads a video to the system, sets the time interval for conversion and submits the video Use case 2 The system processes the video. Precondition:Use case 1 User: System Post condition:the final image The system receives a video from the user, processes it and produces the resultant panoramic image. 5

18 6

19 Chapter 4 Requirements 4.1 Introduction Below are the General user requirements of the System Functional Requirements The System must convert a video into a 3 dimensional panoramic image and it must do this efficiently and accurately. The system must be automated as much as possible. The only part of the system that will not be automated is the one where the user uploads the video and sets the time limits Non-Functional Requirements The system must have a user interface that is easy to use and hides all the complex processes happening in the background. 4.2 Use Case Diagram 7

20 8

21 Chapter 5 User Interface Specification 5.1 Introduction This chapter specifically describes the user interface of the video to panorama converter. 5.2 Description The V2PIC system will be a very simple one in the eyes of a user. Since it s goal is to convert an video into a panoramic image, the system is as simple as having one button to convert the specified video to a panorama. 5.3 User Interface Behavior The user must first interact with the interface by providing the path to the video that is to be converted. This can be done by either specifying the path in typing it out into a field in the interface or by interacting with the interface using a mouse and point to where the video is situated. The user then presses the convert button which will trigger the processing and require the user to wait for some time, preferably a short period of time. The user can then view the resultant panorama. 9

22 10

23 Chapter 6 High Level Design 6.1 Introduction This chapter explains the architecture of the system, identifying all the subsystems that form the complete system. 6.2 Technical Solution Break Down in Subsystems The system has 3 main subsystem, the input, processing and output subsystem. This systematic break down will attempt to give a non-technical explanation of how the system works per component. 6.3 Subsystems descriptions This section describes each subsystem and explains what the function is per subsystem Input This component is a simple component that accepts input as the path of the video and passes it to the processing component Processing The processing component on it s own, consists of two components. The first component is one that detects if the video contains a panorama and if so, picks the key frames that will be stitched together. The second subsystem is the one that stitches frames picked by the detector subsystem Output This subsystem is where the resultant panoramic image ends up. This system will help the user view the panorama in an easily navigable way. 11

24 12

25 Chapter 7 Low Level Design 7.1 Introduction This chapter describes and explains each module of the system from a developer s point of view. The description will go into technical details about what each subsystem must do in order to accomplish it s task. 7.2 Programming language and Libraries The programming language used for the system is python3 including the libraries OpenCV and numpy which are available for python3. OpenCV (Open source Computer Vision Library) is used for certain components of the system including reading the video and extracting each frame for analysis. Numpy is used for the computations involved in each process as it is a powerful library for performing matrix operations which is highly required by the system. 7.3 Modules Description Input The first module of the system which is the input component is the most trivial one.it accepts the path of the video and validates if the provided path is correct or with the help of a regular expression and considering that the video exists in the specified path. When all is done as required, the video must then be passed to the processing module Processing The processing module is broken down into two subsystems. The first subsystem is the one which is responsible for selecting the required frames for the stitching processes. The second module is where the image stitching algorithms are apply. The picking of these frames will be based on the frame rate 13

26 14 of the video thus far because the video will be taken in a controlled manner. The stitching subsystem uses the Gaussian smoothing, SIFT and RANSAC algorithms Gaussian smoothing and Differencing This algorithm is used for extracting features of interest. Gaussian smoothing works by taking a square matrix of some chosen size (kernel size), that consists of elements that are numbers that are normally distributed. The matrix (Gaussian kernel) is convolved with the image (Provided that the image is represented by a matrix of 8 bit pixel values) to produce a new image that consists of a smoothed texture with reduced amount of noise compared to the original one. This technique is applied recursively to each resultant image and the result is stored each time. This is repeated for different scales of the image. Differencing the resultant images and finding the extrema gives key-points for features of interest SIFT SIFT, which stands for Scale Invariant Feature Transform helps to describe features from the key-points obtained from the Gaussian smoothing.it does this by selecting some fixed area of pixels surrounding the key point. The output of this algorithm is a four tuple (p, s, r, f) where p is the pixel co-ordinate,

27 15 s is the scale, r is the orientation/direction and f is the feature descriptor. (Ahi, 2016)Below are the steps for the SIFT algorithm: 1. Compute image gradients in the local 16x16 area at a selected scale. 2. Create an array of orientation/direction histograms using the gradients orientations x 4x4 histogram array = 128 dimensions of our feature descriptor. Below is a figure that illustrates orientation calculation for the descriptor: Figure 7.1: Gradient distribution around key-point. The feature descriptor f is view point independant hence it is the same across multiple images if the key-points indicate the same object in the image RANSAC The RANSAC algorithm makes sure that the feature that are the ones that belong in the transformation from image A to image B. In this way RANSAC rolls out all the outliers to make sure that we have the correct features for an overlap of two images. It does this by creating a model (line) for fitting the maximum amount of inliers from the given data points.a threshold for inliers is set and the distant from each point is computed using squared distances from the line to each point. If a point false outside the threshold, then the point is labeled as an outlier. (Fiscler, 1905) Below are the steps for the RANSAC algorithm: 1. Randomly Sample the number of data points required to fit the model

28 16 2. Comp-ute model parameters using the sample. 3. Score the fraction of inliers within a present threshold of the model. 4. Repeat until model is found. Figure 7.2: RANSAC with inliers and outliers illustration Stitching the images Using the RANSAC algorithm, given Image A and Image B, extracted features from A are compared with the ones from B. The comparison of these features determines which ones overlap (are inliers) and which ones do not (outliers). The features that overlap belong in our transformation model hypothesis. We repeat this process until with the number of repetitions depending on the probability at which we have at least one outlier.we then stitching the Images together by matching the key-points (pixel co-ordinate) where the features are situated according to the feature descriptors. The last step is to stitch each selected frame to the each resultant panorama. This is done as an iterative process until all frames are exhausted. Below is a rough overall algorithm for the stitching.

29 17 Data: Video Result: Panaroma prevframe = Video.Cputure; while Video still playing do set frame rate; currframe = frame from video capture; keypointa = GaussianDiff(prevFrame); keypointb = GaussianDiff(currFrame); descriptora = SIFT(keyPointA); descriptorb = SIFT(keyPointB); matchedset = Matcher(DescriptorA,DescriptorB); resultantimage = Stitcher(matchedSet); prevframe = resultantimage; end Algorithm 1: Algorithm for the stitching process Figure 7.3: Key-point Matching illustration.

30 Object Oriented Design Below is thus far the object oriented structure of how all the modules of the system will work. Figure 7.4: Object Oriented Design

31 Bibliography Ahi, K. (2016). Modeling of terahertz images based on x-ray images: a novel approach for verification of terahertz images and identification of objects with fine details beyond terahertz resolution. Conference Paper, page 3. Fiscler, M. A. (1905). Random Sampling Consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Technical Note, (213):3. Mann, S. and Picard., R. W. (1994). Virtual bellows: constructing highquality images from video., in proceedings of the ieee first international conference on image processingaustin. Technical report. R.P.J. Coset, M.C.G. Leijten, T. M. J. M. (2006). User requirements document,spingrid. Technical report, Technische Informatica, Eindhoven University of Technology. 19

Panoramic Image Stitching

Panoramic Image Stitching Mcgill University Panoramic Image Stitching by Kai Wang Pengbo Li A report submitted in fulfillment for the COMP 558 Final project in the Faculty of Computer Science April 2013 Mcgill University Abstract

More information

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy BSB663 Image Processing Pinar Duygulu Slides are adapted from Selim Aksoy Image matching Image matching is a fundamental aspect of many problems in computer vision. Object or scene recognition Solving

More information

Midterm Examination CS 534: Computational Photography

Midterm Examination CS 534: Computational Photography Midterm Examination CS 534: Computational Photography November 3, 2016 NAME: Problem Score Max Score 1 6 2 8 3 9 4 12 5 4 6 13 7 7 8 6 9 9 10 6 11 14 12 6 Total 100 1 of 8 1. [6] (a) [3] What camera setting(s)

More information

Feature Matching and RANSAC

Feature Matching and RANSAC Feature Matching and RANSAC Recognising Panoramas. [M. Brown and D. Lowe,ICCV 2003] [Brown, Szeliski, Winder, CVPR 2005] with a lot of slides stolen from Steve Seitz, Rick Szeliski, A. Efros Introduction

More information

Local Feature Detectors

Local Feature Detectors Local Feature Detectors Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Slides adapted from Cordelia Schmid and David Lowe, CVPR 2003 Tutorial, Matthew Brown,

More information

Homographies and RANSAC

Homographies and RANSAC Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position

More information

Uncertainties: Representation and Propagation & Line Extraction from Range data

Uncertainties: Representation and Propagation & Line Extraction from Range data 41 Uncertainties: Representation and Propagation & Line Extraction from Range data 42 Uncertainty Representation Section 4.1.3 of the book Sensing in the real world is always uncertain How can uncertainty

More information

Feature Based Registration - Image Alignment

Feature Based Registration - Image Alignment Feature Based Registration - Image Alignment Image Registration Image registration is the process of estimating an optimal transformation between two or more images. Many slides from Alexei Efros http://graphics.cs.cmu.edu/courses/15-463/2007_fall/463.html

More information

Computer Vision. Exercise 3 Panorama Stitching 09/12/2013. Compute Vision : Exercise 3 Panorama Stitching

Computer Vision. Exercise 3 Panorama Stitching 09/12/2013. Compute Vision : Exercise 3 Panorama Stitching Computer Vision Exercise 3 Panorama Stitching 09/12/2013 Compute Vision : Exercise 3 Panorama Stitching The task Compute Vision : Exercise 3 Panorama Stitching 09/12/2013 2 Pipeline Compute Vision : Exercise

More information

CSE 527: Introduction to Computer Vision

CSE 527: Introduction to Computer Vision CSE 527: Introduction to Computer Vision Week 5 - Class 1: Matching, Stitching, Registration September 26th, 2017 ??? Recap Today Feature Matching Image Alignment Panoramas HW2! Feature Matches Feature

More information

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image

More information

Chapter 3 Image Registration. Chapter 3 Image Registration

Chapter 3 Image Registration. Chapter 3 Image Registration Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation

More information

Observations. Basic iteration Line estimated from 2 inliers

Observations. Basic iteration Line estimated from 2 inliers Line estimated from 2 inliers 3 Observations We need (in this case!) a minimum of 2 points to determine a line Given such a line l, we can determine how well any other point y fits the line l For example:

More information

Image Stitching using Watersheds and Graph Cuts

Image Stitching using Watersheds and Graph Cuts Image Stitching using Watersheds and Graph Cuts Patrik Nyman Centre for Mathematical Sciences, Lund University, Sweden patnym@maths.lth.se 1. Introduction Image stitching is commonly used in many different

More information

Image stitching. Announcements. Outline. Image stitching

Image stitching. Announcements. Outline. Image stitching Announcements Image stitching Project #1 was due yesterday. Project #2 handout will be available on the web later tomorrow. I will set up a webpage for artifact voting soon. Digital Visual Effects, Spring

More information

From Structure-from-Motion Point Clouds to Fast Location Recognition

From Structure-from-Motion Point Clouds to Fast Location Recognition From Structure-from-Motion Point Clouds to Fast Location Recognition Arnold Irschara1;2, Christopher Zach2, Jan-Michael Frahm2, Horst Bischof1 1Graz University of Technology firschara, bischofg@icg.tugraz.at

More information

Mosaics. Today s Readings

Mosaics. Today s Readings Mosaics VR Seattle: http://www.vrseattle.com/ Full screen panoramas (cubic): http://www.panoramas.dk/ Mars: http://www.panoramas.dk/fullscreen3/f2_mars97.html Today s Readings Szeliski and Shum paper (sections

More information

Location Based Selling Platform for Mobile Buyers

Location Based Selling Platform for Mobile Buyers Location Based Selling Platform for Mobile Buyers M. M. Buddhika Mawella 149219M Faculty of Information Technology University of Moratuwa April 2017 Location Based Selling Platform for Mobile Buyers M.

More information

Real Time Multi-Sensor Data Acquisition and Processing for a Road Mapping System

Real Time Multi-Sensor Data Acquisition and Processing for a Road Mapping System Real Time Multi-Sensor Data Acquisition and Processing for a Road Mapping System by Xiang Luo A thesis submitted for the degree of Master of Engineering (Research) Faculty of Engineering and Information

More information

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58

Image Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 Image Features: Local Descriptors Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 [Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 2/ 58 Local Features Detection: Identify

More information

Advanced Video Content Analysis and Video Compression (5LSH0), Module 4

Advanced Video Content Analysis and Video Compression (5LSH0), Module 4 Advanced Video Content Analysis and Video Compression (5LSH0), Module 4 Visual feature extraction Part I: Color and texture analysis Sveta Zinger Video Coding and Architectures Research group, TU/e ( s.zinger@tue.nl

More information

Image correspondences and structure from motion

Image correspondences and structure from motion Image correspondences and structure from motion http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 20 Course announcements Homework 5 posted.

More information

Analysis of Image and Video Using Color, Texture and Shape Features for Object Identification

Analysis of Image and Video Using Color, Texture and Shape Features for Object Identification IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 16, Issue 6, Ver. VI (Nov Dec. 2014), PP 29-33 Analysis of Image and Video Using Color, Texture and Shape Features

More information

Robotics Programming Laboratory

Robotics Programming Laboratory Chair of Software Engineering Robotics Programming Laboratory Bertrand Meyer Jiwon Shin Lecture 8: Robot Perception Perception http://pascallin.ecs.soton.ac.uk/challenges/voc/databases.html#caltech car

More information

A Comparison of SIFT, PCA-SIFT and SURF

A Comparison of SIFT, PCA-SIFT and SURF A Comparison of SIFT, PCA-SIFT and SURF Luo Juan Computer Graphics Lab, Chonbuk National University, Jeonju 561-756, South Korea qiuhehappy@hotmail.com Oubong Gwun Computer Graphics Lab, Chonbuk National

More information

CS1114: Study Guide 2

CS1114: Study Guide 2 CS4: Study Guide 2 This document covers the topics we ve covered in the second part of the course. Please refer to the class slides for more details. Polygons and convex hulls A polygon is a set of 2D

More information

Hough Transform and RANSAC

Hough Transform and RANSAC CS4501: Introduction to Computer Vision Hough Transform and RANSAC Various slides from previous courses by: D.A. Forsyth (Berkeley / UIUC), I. Kokkinos (Ecole Centrale / UCL). S. Lazebnik (UNC / UIUC),

More information

Perception IV: Place Recognition, Line Extraction

Perception IV: Place Recognition, Line Extraction Perception IV: Place Recognition, Line Extraction Davide Scaramuzza University of Zurich Margarita Chli, Paul Furgale, Marco Hutter, Roland Siegwart 1 Outline of Today s lecture Place recognition using

More information

Object Recognition with Invariant Features

Object Recognition with Invariant Features Object Recognition with Invariant Features Definition: Identify objects or scenes and determine their pose and model parameters Applications Industrial automation and inspection Mobile robots, toys, user

More information

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Stitching = alignment + blending geometrical registration

More information

Local Features Tutorial: Nov. 8, 04

Local Features Tutorial: Nov. 8, 04 Local Features Tutorial: Nov. 8, 04 Local Features Tutorial References: Matlab SIFT tutorial (from course webpage) Lowe, David G. Distinctive Image Features from Scale Invariant Features, International

More information

Local invariant features

Local invariant features Local invariant features Tuesday, Oct 28 Kristen Grauman UT-Austin Today Some more Pset 2 results Pset 2 returned, pick up solutions Pset 3 is posted, due 11/11 Local invariant features Detection of interest

More information

A Summary of Projective Geometry

A Summary of Projective Geometry A Summary of Projective Geometry Copyright 22 Acuity Technologies Inc. In the last years a unified approach to creating D models from multiple images has been developed by Beardsley[],Hartley[4,5,9],Torr[,6]

More information

Scale Invariant Feature Transform

Scale Invariant Feature Transform Scale Invariant Feature Transform Why do we care about matching features? Camera calibration Stereo Tracking/SFM Image moiaicing Object/activity Recognition Objection representation and recognition Image

More information

Department of Computer Science University of Western Cape Computer Science Honours Project: Anaglyph Video By: Jihaad Pienaar Supervisors: Mehrdad

Department of Computer Science University of Western Cape Computer Science Honours Project: Anaglyph Video By: Jihaad Pienaar Supervisors: Mehrdad Department of Computer Science University of Western Cape Computer Science Honours Project: Anaglyph Video By: Jihaad Pienaar Supervisors: Mehrdad Ghaziasgar and James Connan Glossary Depth Map Stereo

More information

Panoramic Image Mosaicing

Panoramic Image Mosaicing Panoramic Image Mosaicing Mrs. Hetal M.Patel 1, Asst. Prof. Pinal. J. Patel 2, Asst. Prof. Mr. Sandip G. Patel 3 1 C.S.E. Department, Government College of Engineering, Gandhinagar Gujarat Technology University,

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION

A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM INTRODUCTION A NEW FEATURE BASED IMAGE REGISTRATION ALGORITHM Karthik Krish Stuart Heinrich Wesley E. Snyder Halil Cakir Siamak Khorram North Carolina State University Raleigh, 27695 kkrish@ncsu.edu sbheinri@ncsu.edu

More information

Multi-stable Perception. Necker Cube

Multi-stable Perception. Necker Cube Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuyuki Kayahara Multiple view geometry Stereo vision Epipolar geometry Lowe Hartley and Zisserman Depth map extraction Essential matrix

More information

Object Reconstruction

Object Reconstruction B. Scholz Object Reconstruction 1 / 39 MIN-Fakultät Fachbereich Informatik Object Reconstruction Benjamin Scholz Universität Hamburg Fakultät für Mathematik, Informatik und Naturwissenschaften Fachbereich

More information

[Programming Assignment] (1)

[Programming Assignment] (1) http://crcv.ucf.edu/people/faculty/bagci/ [Programming Assignment] (1) Computer Vision Dr. Ulas Bagci (Fall) 2015 University of Central Florida (UCF) Coding Standard and General Requirements Code for all

More information

Feature Detectors and Descriptors: Corners, Lines, etc.

Feature Detectors and Descriptors: Corners, Lines, etc. Feature Detectors and Descriptors: Corners, Lines, etc. Edges vs. Corners Edges = maxima in intensity gradient Edges vs. Corners Corners = lots of variation in direction of gradient in a small neighborhood

More information

Perception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich.

Perception. Autonomous Mobile Robots. Sensors Vision Uncertainties, Line extraction from laser scans. Autonomous Systems Lab. Zürich. Autonomous Mobile Robots Localization "Position" Global Map Cognition Environment Model Local Map Path Perception Real World Environment Motion Control Perception Sensors Vision Uncertainties, Line extraction

More information

A System of Image Matching and 3D Reconstruction

A System of Image Matching and 3D Reconstruction A System of Image Matching and 3D Reconstruction CS231A Project Report 1. Introduction Xianfeng Rui Given thousands of unordered images of photos with a variety of scenes in your gallery, you will find

More information

Global Flow Estimation. Lecture 9

Global Flow Estimation. Lecture 9 Motion Models Image Transformations to relate two images 3D Rigid motion Perspective & Orthographic Transformation Planar Scene Assumption Transformations Translation Rotation Rigid Affine Homography Pseudo

More information

The SIFT (Scale Invariant Feature

The SIFT (Scale Invariant Feature The SIFT (Scale Invariant Feature Transform) Detector and Descriptor developed by David Lowe University of British Columbia Initial paper ICCV 1999 Newer journal paper IJCV 2004 Review: Matt Brown s Canonical

More information

Prof. Noah Snavely CS Administrivia. A4 due on Friday (please sign up for demo slots)

Prof. Noah Snavely CS Administrivia. A4 due on Friday (please sign up for demo slots) Robust fitting Prof. Noah Snavely CS111 http://www.cs.cornell.edu/courses/cs111 Administrivia A due on Friday (please sign up for demo slots) A5 will be out soon Prelim is coming up, Tuesday, / Roadmap

More information

Automatic Image Alignment (feature-based)

Automatic Image Alignment (feature-based) Automatic Image Alignment (feature-based) Mike Nese with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2006 Today s lecture Feature

More information

URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES

URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES An Undergraduate Research Scholars Thesis by RUI LIU Submitted to Honors and Undergraduate Research Texas A&M University in partial fulfillment

More information

Scale Invariant Feature Transform

Scale Invariant Feature Transform Why do we care about matching features? Scale Invariant Feature Transform Camera calibration Stereo Tracking/SFM Image moiaicing Object/activity Recognition Objection representation and recognition Automatic

More information

Computer Vision, Assignment 4 Model Fitting

Computer Vision, Assignment 4 Model Fitting Centre for Mathematical Sciences, February 2013 Due 2013-02-26 Computer Vision, Assignment 4 Model Fitting 1 Instructions In this assignment you will study model fitting. In particular you will use random

More information

Stitching and Blending

Stitching and Blending Stitching and Blending Kari Pulli VP Computational Imaging Light First project Build your own (basic) programs panorama HDR (really, exposure fusion) The key components register images so their features

More information

CLASSIFICATION AND CHANGE DETECTION

CLASSIFICATION AND CHANGE DETECTION IMAGE ANALYSIS, CLASSIFICATION AND CHANGE DETECTION IN REMOTE SENSING With Algorithms for ENVI/IDL and Python THIRD EDITION Morton J. Canty CRC Press Taylor & Francis Group Boca Raton London NewYork CRC

More information

CSE 252B: Computer Vision II

CSE 252B: Computer Vision II CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribes: Jeremy Pollock and Neil Alldrin LECTURE 14 Robust Feature Matching 14.1. Introduction Last lecture we learned how to find interest points

More information

Comparison of Feature Detection and Matching Approaches: SIFT and SURF

Comparison of Feature Detection and Matching Approaches: SIFT and SURF GRD Journals- Global Research and Development Journal for Engineering Volume 2 Issue 4 March 2017 ISSN: 2455-5703 Comparison of Detection and Matching Approaches: SIFT and SURF Darshana Mistry PhD student

More information

CS 558: Computer Vision 4 th Set of Notes

CS 558: Computer Vision 4 th Set of Notes 1 CS 558: Computer Vision 4 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Overview Keypoint matching Hessian

More information

Digital Image Processing

Digital Image Processing Digital Image Processing Third Edition Rafael C. Gonzalez University of Tennessee Richard E. Woods MedData Interactive PEARSON Prentice Hall Pearson Education International Contents Preface xv Acknowledgments

More information

Automatic Image Alignment

Automatic Image Alignment Automatic Image Alignment Mike Nese with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2010 Live Homography DEMO Check out panoramio.com

More information

Introduction to Medical Imaging (5XSA0)

Introduction to Medical Imaging (5XSA0) 1 Introduction to Medical Imaging (5XSA0) Visual feature extraction Color and texture analysis Sveta Zinger ( s.zinger@tue.nl ) Introduction (1) Features What are features? Feature a piece of information

More information

Gesture based PTZ camera control

Gesture based PTZ camera control Gesture based PTZ camera control Report submitted in May 2014 to the department of Computer Science and Engineering of National Institute of Technology Rourkela in partial fulfillment of the requirements

More information

INFO0948 Fitting and Shape Matching

INFO0948 Fitting and Shape Matching INFO0948 Fitting and Shape Matching Renaud Detry University of Liège, Belgium Updated March 31, 2015 1 / 33 These slides are based on the following book: D. Forsyth and J. Ponce. Computer vision: a modern

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

Model Fitting, RANSAC. Jana Kosecka

Model Fitting, RANSAC. Jana Kosecka Model Fitting, RANSAC Jana Kosecka Fitting: Overview If we know which points belong to the line, how do we find the optimal line parameters? Least squares What if there are outliers? Robust fitting, RANSAC

More information

521466S Machine Vision Assignment #3 Image Features

521466S Machine Vision Assignment #3 Image Features 521466S Machine Vision Assignment #3 Image Features Spring 2018 This assignment explores feature detection, extraction, and matching. We will implement the well-known Harris corner detector and use Matlab

More information

Implementing the Scale Invariant Feature Transform(SIFT) Method

Implementing the Scale Invariant Feature Transform(SIFT) Method Implementing the Scale Invariant Feature Transform(SIFT) Method YU MENG and Dr. Bernard Tiddeman(supervisor) Department of Computer Science University of St. Andrews yumeng@dcs.st-and.ac.uk Abstract The

More information

An Algorithm for Seamless Image Stitching and Its Application

An Algorithm for Seamless Image Stitching and Its Application An Algorithm for Seamless Image Stitching and Its Application Jing Xing, Zhenjiang Miao, and Jing Chen Institute of Information Science, Beijing JiaoTong University, Beijing 100044, P.R. China Abstract.

More information

A Keypoint Descriptor Inspired by Retinal Computation

A Keypoint Descriptor Inspired by Retinal Computation A Keypoint Descriptor Inspired by Retinal Computation Bongsoo Suh, Sungjoon Choi, Han Lee Stanford University {bssuh,sungjoonchoi,hanlee}@stanford.edu Abstract. The main goal of our project is to implement

More information

Instance-level recognition

Instance-level recognition Instance-level recognition 1) Local invariant features 2) Matching and recognition with local features 3) Efficient visual search 4) Very large scale indexing Matching of descriptors Matching and 3D reconstruction

More information

CS 378: Autonomous Intelligent Robotics. Instructor: Jivko Sinapov

CS 378: Autonomous Intelligent Robotics. Instructor: Jivko Sinapov CS 378: Autonomous Intelligent Robotics Instructor: Jivko Sinapov http://www.cs.utexas.edu/~jsinapov/teaching/cs378/ Visual Registration and Recognition Announcements Homework 6 is out, due 4/5 4/7 Installing

More information

School of Computing University of Utah

School of Computing University of Utah School of Computing University of Utah Presentation Outline 1 2 3 4 Main paper to be discussed David G. Lowe, Distinctive Image Features from Scale-Invariant Keypoints, IJCV, 2004. How to find useful keypoints?

More information

Image Segmentation and Registration

Image Segmentation and Registration Image Segmentation and Registration Dr. Christine Tanner (tanner@vision.ee.ethz.ch) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation

More information

RANSAC and some HOUGH transform

RANSAC and some HOUGH transform RANSAC and some HOUGH transform Thank you for the slides. They come mostly from the following source Dan Huttenlocher Cornell U Matching and Fitting Recognition and matching are closely related to fitting

More information

A Novel Extreme Point Selection Algorithm in SIFT

A Novel Extreme Point Selection Algorithm in SIFT A Novel Extreme Point Selection Algorithm in SIFT Ding Zuchun School of Electronic and Communication, South China University of Technolog Guangzhou, China zucding@gmail.com Abstract. This paper proposes

More information

Motivation. Intensity Levels

Motivation. Intensity Levels Motivation Image Intensity and Point Operations Dr. Edmund Lam Department of Electrical and Electronic Engineering The University of Hong ong A digital image is a matrix of numbers, each corresponding

More information

3D OBJECT RECONTRUCTION USING MULTIPLE-VIEW GEOMETRY: SIFT DETECTION LEOW RUEY SHYAN

3D OBJECT RECONTRUCTION USING MULTIPLE-VIEW GEOMETRY: SIFT DETECTION LEOW RUEY SHYAN 3D OBJECT RECONTRUCTION USING MULTIPLE-VIEW GEOMETRY: SIFT DETECTION LEOW RUEY SHYAN A project report submitted in partial fulfilment of the requirements for the award of the degree of Bachelor (Hons.)

More information

UNIVERSITI MALAYSIA PAHANG

UNIVERSITI MALAYSIA PAHANG IMAGE ENHANCEMENT AND SEGMENTATION ON SIMULTANEOUS LATENT FINGERPRINT DETECTION ROZITA BINTI MOHD YUSOF MASTER OF COMPUTER SCIENCE UNIVERSITI MALAYSIA PAHANG IMAGE ENHANCEMENT AND SEGMENTATION ON SIMULTANEOUS

More information

Combining Appearance and Topology for Wide

Combining Appearance and Topology for Wide Combining Appearance and Topology for Wide Baseline Matching Dennis Tell and Stefan Carlsson Presented by: Josh Wills Image Point Correspondences Critical foundation for many vision applications 3-D reconstruction,

More information

Accurate Image Registration from Local Phase Information

Accurate Image Registration from Local Phase Information Accurate Image Registration from Local Phase Information Himanshu Arora, Anoop M. Namboodiri, and C.V. Jawahar Center for Visual Information Technology, IIIT, Hyderabad, India { himanshu@research., anoop@,

More information

Miniature faking. In close-up photo, the depth of field is limited.

Miniature faking. In close-up photo, the depth of field is limited. Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg

More information

Face and Nose Detection in Digital Images using Local Binary Patterns

Face and Nose Detection in Digital Images using Local Binary Patterns Face and Nose Detection in Digital Images using Local Binary Patterns Stanko Kružić Post-graduate student University of Split, Faculty of Electrical Engineering, Mechanical Engineering and Naval Architecture

More information

The Lucas & Kanade Algorithm

The Lucas & Kanade Algorithm The Lucas & Kanade Algorithm Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Registration, Registration, Registration. Linearizing Registration. Lucas & Kanade Algorithm. 3 Biggest

More information

Fundamentals of Digital Image Processing

Fundamentals of Digital Image Processing \L\.6 Gw.i Fundamentals of Digital Image Processing A Practical Approach with Examples in Matlab Chris Solomon School of Physical Sciences, University of Kent, Canterbury, UK Toby Breckon School of Engineering,

More information

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Stitching = alignment + blending geometrical registration

More information

3D Environment Reconstruction

3D Environment Reconstruction 3D Environment Reconstruction Using Modified Color ICP Algorithm by Fusion of a Camera and a 3D Laser Range Finder The 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems October 11-15,

More information

Computer Vision I - Filtering and Feature detection

Computer Vision I - Filtering and Feature detection Computer Vision I - Filtering and Feature detection Carsten Rother 30/10/2015 Computer Vision I: Basics of Image Processing Roadmap: Basics of Digital Image Processing Computer Vision I: Basics of Image

More information

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah School of Computing University of Utah Presentation Outline 1 2 3 Forward Projection (Reminder) u v 1 KR ( I t ) X m Y m Z m 1 Backward Projection (Reminder) Q K 1 q Presentation Outline 1 2 3 Sample Problem

More information

Computer Vision CSCI-GA Assignment 1.

Computer Vision CSCI-GA Assignment 1. Computer Vision CSCI-GA.2272-001 Assignment 1. September 22, 2017 Introduction This assignment explores various methods for aligning images and feature extraction. There are four parts to the assignment:

More information

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS Cognitive Robotics Original: David G. Lowe, 004 Summary: Coen van Leeuwen, s1460919 Abstract: This article presents a method to extract

More information

Feature Matching and Robust Fitting

Feature Matching and Robust Fitting Feature Matching and Robust Fitting Computer Vision CS 143, Brown Read Szeliski 4.1 James Hays Acknowledgment: Many slides from Derek Hoiem and Grauman&Leibe 2008 AAAI Tutorial Project 2 questions? This

More information

Computer Vision for HCI. Topics of This Lecture

Computer Vision for HCI. Topics of This Lecture Computer Vision for HCI Interest Points Topics of This Lecture Local Invariant Features Motivation Requirements, Invariances Keypoint Localization Features from Accelerated Segment Test (FAST) Harris Shi-Tomasi

More information

EVALUATION OF SEQUENTIAL IMAGES FOR PHOTOGRAMMETRICALLY POINT DETERMINATION

EVALUATION OF SEQUENTIAL IMAGES FOR PHOTOGRAMMETRICALLY POINT DETERMINATION Archives of Photogrammetry, Cartography and Remote Sensing, Vol. 22, 2011, pp. 285-296 ISSN 2083-2214 EVALUATION OF SEQUENTIAL IMAGES FOR PHOTOGRAMMETRICALLY POINT DETERMINATION Michał Kowalczyk 1 1 Department

More information

Obtaining Feature Correspondences

Obtaining Feature Correspondences Obtaining Feature Correspondences Neill Campbell May 9, 2008 A state-of-the-art system for finding objects in images has recently been developed by David Lowe. The algorithm is termed the Scale-Invariant

More information

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking

Feature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking Feature descriptors Alain Pagani Prof. Didier Stricker Computer Vision: Object and People Tracking 1 Overview Previous lectures: Feature extraction Today: Gradiant/edge Points (Kanade-Tomasi + Harris)

More information

Local features: detection and description May 12 th, 2015

Local features: detection and description May 12 th, 2015 Local features: detection and description May 12 th, 2015 Yong Jae Lee UC Davis Announcements PS1 grades up on SmartSite PS1 stats: Mean: 83.26 Standard Dev: 28.51 PS2 deadline extended to Saturday, 11:59

More information

Image processing and features

Image processing and features Image processing and features Gabriele Bleser gabriele.bleser@dfki.de Thanks to Harald Wuest, Folker Wientapper and Marc Pollefeys Introduction Previous lectures: geometry Pose estimation Epipolar geometry

More information

An Implementation on Histogram of Oriented Gradients for Human Detection

An Implementation on Histogram of Oriented Gradients for Human Detection An Implementation on Histogram of Oriented Gradients for Human Detection Cansın Yıldız Dept. of Computer Engineering Bilkent University Ankara,Turkey cansin@cs.bilkent.edu.tr Abstract I implemented a Histogram

More information

CITS 4402 Computer Vision

CITS 4402 Computer Vision CITS 4402 Computer Vision A/Prof Ajmal Mian Adj/A/Prof Mehdi Ravanbakhsh, CEO at Mapizy (www.mapizy.com) and InFarm (www.infarm.io) Lecture 02 Binary Image Analysis Objectives Revision of image formation

More information

CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry

CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry Some material taken from: David Lowe, UBC Jiri Matas, CMP Prague http://cmp.felk.cvut.cz/~matas/papers/presentations/matas_beyondransac_cvprac05.ppt

More information

TA Section 7 Problem Set 3. SIFT (Lowe 2004) Shape Context (Belongie et al. 2002) Voxel Coloring (Seitz and Dyer 1999)

TA Section 7 Problem Set 3. SIFT (Lowe 2004) Shape Context (Belongie et al. 2002) Voxel Coloring (Seitz and Dyer 1999) TA Section 7 Problem Set 3 SIFT (Lowe 2004) Shape Context (Belongie et al. 2002) Voxel Coloring (Seitz and Dyer 1999) Sam Corbett-Davies TA Section 7 02-13-2014 Distinctive Image Features from Scale-Invariant

More information

Classification of objects from Video Data (Group 30)

Classification of objects from Video Data (Group 30) Classification of objects from Video Data (Group 30) Sheallika Singh 12665 Vibhuti Mahajan 12792 Aahitagni Mukherjee 12001 M Arvind 12385 1 Motivation Video surveillance has been employed for a long time

More information