Camera Calibration Utility Description

Size: px
Start display at page:

Download "Camera Calibration Utility Description"

Transcription

1 Camera Calibration Utility Description Robert Bryll, Xinfeng Ma, Francis Quek Vision Interfaces and Systems Laboratory The university of Illinois at Chicago April 6, Introduction To calibrate our cameras for stereo viewing we use Roger Y. Tsai s versatile camera calibration algorithm described in [1]. The problem with calibrating two cameras for stereo viewing is that to calibrate them it is necessary to record a calibration target with points of known coordinates and then find the image coordinates of calibration points on images from both cameras, put them in files in appropriate order and pass them to the calibration routines. Tsai s algorithm requires at least 11 calibration points, but the normally used number is between 20 and 60. Picking those points by hand to create input files for calibration rooutines is tedious and time consuming. We make the proces of finding calibration points in frames captured from both cameras semiautomatic. 2 Solution The calibration target (a plywood box with two walls put at an angle of 120 degrees to one another) is painted white and all the calibration points (of known coordinates) are marked with black 3/4 circles. The size of the calibration target approximates the average volume of the human gesture space. The calibration target is presented in Figure 1. Figure 2 shows how the world coordinate system is positioned with respect to the calibration target. During the calibration the box should fill approx. 1/3 rd of the frame area. Black calibration points are detected semi-automatically by the following steps: 1. Threshold the input images to detect dark areas (thresholds are adjusted by hand). 2. Find connected components in the image, label the detected regions. 3. Run size filter over the detected regions. Since the calibration points are relatively small (around pixels according to our tests), this step filters out many unwanted regions. 4. Remove false positives by hand. 5. Compute centers of gravity of the final set of regions to obtain more precise locations of the calibration points. 1

2 Figure 1: Stereo camera calibration target used in our experiments. 2.1 Camera Calibration Utility (CCU) Figure 3 shows the Camera Calibration Utility interface that allows the user to perform the steps described above semi-automatically. The following sections describe the steps necessary to perform the calibration point detection and to save the resulting files Input Files There are two input files for the calbration (one image from each camera). They should both be in the raw data grayscale.pgm format, both have the same dimensions (close to pixels). The names of the files should follow a simple convention: the filename of the lower image is equal to the name of the upper image with extension.2. So, for example, if the upper image file is calib1.pgm, the lower image file should be calib1.pgm.2. Both files are opened by pressing the Load Images button and selecting the name of the upper image only Thresholding After opening the image files, they are displayed in the window as in the figure 3. Then the user has to adjust the thresholds for both images, so that the calibration points are clearly defined on the faces of the calibration target on both images. The thresholds are adjusted by moving the sliders in the Thresholds part of the interface window. When sliders are moved, the corresponding image changes, showing the result of the thresholding. The faces of the calibration target should be black 2

3 X Z Y X Figure 2: Position of the World Coordinate System with repect to the calibration target. and calibration points should be represented by white dots. The default values of the thresholds work in some cases, but usually they have to be adjusted (or at least the results of the thresholding have to be checked, by clicking on the sliders). It is possible to display the original images at any time by pressing the Show Originals button. It is also possible to adjust size filters ro the calibration points, but in our experinments it was never necessary. It may be necessary if the calibration target is very close or very far away from the cameras Detecting Calibration Points After the thresholds have been set, the user should press the Detect Points button to perform the actual detection. The program filters the image with 55 Gaussian filter, then thresholds it according to the adjusted threshold values, finds connected components and runs size filter over them. Finally, it computes the ceter of gravity of each accepted region and displays the ceners positions on the original images. The accepted regions are represented by red crosses with corresponding numbers. After this stage the regions are unsorted and there is no correlation between the regions on both images Removing False Positives After the detection stage there usually are some false positives that have to be removed. The point can be removed by left-clicking on the red cross representing the calibration point (another click will turn the point back on). All false positives must be removed, and moreover, the sets of points on both calibration images have to be identical (the same number of points, same positions on the target!). For example, in figure 3 we can see that on both images the leftmost column of calibration points is unsused. If on the upper image there are calibration points in this column, and on the lower one there are no calibration points there (because of false negatives), the calibration points have to be removed from the upper image to make both sets identical. 3

4 Figure 3: Camera Calibration Utility Interface. 4

5 2.1.5 Setting the Points on the Calibration Points Panel and Sorting the Calibration Points In order the correspondence between two sets of points to be resolved correctly, the sets of calibration points have to be reflected on the Calibration Points Panel. Calibration Points Panel contains two sets of 25 buttons, each set represents one face of the calibration target. Initially all buttons contain are checked (contain letter x), whicgh means that all 50 points are selected. However, if there are false negatives on the calibration images, the user has to uncheck the necessary calibration points to reflect the actual detected state. Figure 3 shows the situation in which the leftmost column of the calibration points is unchecked, and there are four unused (undetected) calibration points in the center of the calibration target. Setting the Calibration Points Panel is very important for sorting the calibration points on both images so that there is exact correspondence between the two images. The sorting is performed as follows: 1. The calibration points array for each image is first sorted in place with respect to the image X coordinate, so that the numbers of points are increasing in columns starting from the left. The assumption here is that the target is relatively vertical, that is there is no overlap between the calibration points columns in the X direction. If the target in the image is very far from vertical, the algorithm will fail (because it won t be able to distinguish the columns if there is overlap between the points belonging to different columns in the X direction). However, such situation didn t occur in our experiments, so we think the assumption is safe. 2. The calibration points array for each image is then piecewise sorted in place using the information from the Calibration Points Panel. The pieces sorted correspond to the columns in the calibration target and sorting is done with respect to the image Y coordinate (so now the points are sorted only in columns). That s why the correct information from the Calibration Points Panel is crucial for the algorithm (because the numbers of points in columns do not have to be equal). As the result, the points are ordered starting from the upper point in the leftmost column, and the numbers increase when going down in columns and then to the right between columns. The sortingis done after setting the CalibrationPointsPaneland pressing the SORT POINTS button. The program shows the resulting order (calibration points with their numbers). The situation after sorting is presented in figure 3. After sorting there is exact correspondence between the calibration points on both images and their coordinates can be saved for further processing (using Xinfeng s calibration interface as a wrapper over the Tsai s routines) Saving the Calibration Files After sorting the calibration points, they can be saved into calibration files (simple text files) using two buttons in the lower left corner of the interface window. The Save World & Image Coords button saves 5 floating point numbers for each calibration point: first the three world coordinates (x,y,z) of each detected point and then its two image coordinates (computed with subpixel accuracy as the regions center of gravity). The Save Image Coords Only button saves only the two image coordinates for each calibration point on each calibration image. 3 Naming Conventions for Files The names of saved calibration files follow simple naming conventions. We will show them by example. If the fupper calibration image is called calib1.pgm, the lower image has to be called calib1.pgm.2. Then the names of calibration files for both images will be as follows: 5

6 calib1.pgm.world.dat - the file with world and image coordinates for the upper image calib1.pgm.2.world.dat - the file with world and image coordinates for the lower image calib.pgm.image.dat - the file with only image coordinates for the upper image calib.pgm.2.image.dat - the file with only image coordinates for the lower image The files can then be used (together with the camera data files) by the Xinfeng s camera calibration interface built over the Tsai s calibration routines. 4 Accuracy Tests We wrote a simple program analyzing the output from the world data files after 3D triangulation and comparing it to the original world coordinates. We performed the accuracy tests by first calibrating the camera using a set of calibration pictures and then using the same pictures (image coordinates only) to calculate the real world coordinates. Comparing the calculated world coordinates with the original calibration world coordinates yielded the calibration errors in 3 dimensions. Here are the results of our tests (all errors in mm): 1. Baseline Length approximately 1953 mm; 38 calibration points: Average error in X: mm Maximal error in X: mm Average error in Y: mm Maximal error in Y: mm Average error in Z: mm Maximal error in Z: mm 2. Baseline Length Approximately 1198 mm; 43 calibration points: Average error in X: mm Maximal error in X: mm Average error in Y: mm Maximal error in Y: mm Average error in Z: mm Maximal error in Z: mm 3. Baseline Length Approximately 1186 mm; 39 calibration points: Average error in X: mm Maximal error in X: mm Average error in Y: mm Maximal error in Y: mm Average error in Z: mm Maximal error in Z: mm It can easily be observed that the longer baseline (point 1 above) gives higher accuracy of triangulation. However, even for the shorter baselines the maximal triangulation error never exceeds 5mm, which is adequate for our 3D gesture experiments. 6

7 References [1] R.Y. Tsai, A versatile camera calibration technique for high accuracy 3d machine vision metrology using off-the-shelf TV cameras and lenses, IEEE Journal of Robotics and Automation, vol. RA-3, pp ,

PRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1. We present some preliminary results on a system for tracking 3D motion using

PRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1. We present some preliminary results on a system for tracking 3D motion using PRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1 Tak-keung CHENG derek@cs.mu.oz.au Leslie KITCHEN ljk@cs.mu.oz.au Computer Vision and Pattern Recognition Laboratory, Department of Computer Science,

More information

DEVELOPMENT OF REAL TIME 3-D MEASUREMENT SYSTEM USING INTENSITY RATIO METHOD

DEVELOPMENT OF REAL TIME 3-D MEASUREMENT SYSTEM USING INTENSITY RATIO METHOD DEVELOPMENT OF REAL TIME 3-D MEASUREMENT SYSTEM USING INTENSITY RATIO METHOD Takeo MIYASAKA and Kazuo ARAKI Graduate School of Computer and Cognitive Sciences, Chukyo University, Japan miyasaka@grad.sccs.chukto-u.ac.jp,

More information

EE368 Project: Visual Code Marker Detection

EE368 Project: Visual Code Marker Detection EE368 Project: Visual Code Marker Detection Kahye Song Group Number: 42 Email: kahye@stanford.edu Abstract A visual marker detection algorithm has been implemented and tested with twelve training images.

More information

Bumblebee2 Stereo Vision Camera

Bumblebee2 Stereo Vision Camera Bumblebee2 Stereo Vision Camera Description We use the Point Grey Bumblebee2 Stereo Vision Camera in this lab section. This stereo camera can capture 648 x 488 video at 48 FPS. 1) Microlenses 2) Status

More information

HIGH SPEED 3-D MEASUREMENT SYSTEM USING INCOHERENT LIGHT SOURCE FOR HUMAN PERFORMANCE ANALYSIS

HIGH SPEED 3-D MEASUREMENT SYSTEM USING INCOHERENT LIGHT SOURCE FOR HUMAN PERFORMANCE ANALYSIS HIGH SPEED 3-D MEASUREMENT SYSTEM USING INCOHERENT LIGHT SOURCE FOR HUMAN PERFORMANCE ANALYSIS Takeo MIYASAKA, Kazuhiro KURODA, Makoto HIROSE and Kazuo ARAKI School of Computer and Cognitive Sciences,

More information

A High Speed Face Measurement System

A High Speed Face Measurement System A High Speed Face Measurement System Kazuhide HASEGAWA, Kazuyuki HATTORI and Yukio SATO Department of Electrical and Computer Engineering, Nagoya Institute of Technology Gokiso, Showa, Nagoya, Japan, 466-8555

More information

Eagle Knights: RoboCup Small Size League. Dr. Alfredo Weitzenfeld ITAM - Mexico

Eagle Knights: RoboCup Small Size League. Dr. Alfredo Weitzenfeld ITAM - Mexico Eagle Knights: RoboCup Small Size League Dr. Alfredo Weitzenfeld ITAM - Mexico Playing Field and Robot Size camera 15 cm 4 m 18 cm 4 m 5.5 m Distributed Robot Control Eagle Knights (2003-2006) 2003 2004

More information

SRI Small Vision System

SRI Small Vision System Small Vision System Calibration 1 SRI Small Vision System Calibration Supplement to the User s Manual Software version 2.2b July 2001 Kurt Konolige and David Beymer SRI International konolige@ai.sri.com

More information

A topological approach to finding grids in calibration patterns

A topological approach to finding grids in calibration patterns Machine Vision and Applications (2010) 21:949 957 DOI 10.1007/s00138-009-0202-2 SHORT PAPER A topological approach to finding grids in calibration patterns Chang Shu Alan Brunton Mark Fiala Received: 15

More information

ios App User Manual Fill in the serial number (case insensitive), login (used to recover password), and choose a password.

ios App User Manual Fill in the serial number (case insensitive), login  (used to recover password), and choose a password. 1 ios App User Manual Install our iphone/ipad app from AppStore by searching "Wireless Tag List". Use the 12 digit serial numbers on your Ethernet Tag Manager to create a login. Keep the serial number

More information

Infrared Camera Calibration in the 3D Temperature Field Reconstruction

Infrared Camera Calibration in the 3D Temperature Field Reconstruction , pp.27-34 http://dx.doi.org/10.14257/ijmue.2016.11.6.03 Infrared Camera Calibration in the 3D Temperature Field Reconstruction Sun Xiaoming, Wu Haibin, Wang Wei, Liubo and Cui Guoguang The higher Educational

More information

3D-OBJECT DETECTION METHOD BASED ON THE STEREO IMAGE TRANSFORMATION TO THE COMMON OBSERVATION POINT

3D-OBJECT DETECTION METHOD BASED ON THE STEREO IMAGE TRANSFORMATION TO THE COMMON OBSERVATION POINT 3D-OBJECT DETECTION METHOD BASED ON THE STEREO IMAGE TRANSFORMATION TO THE COMMON OBSERVATION POINT V. M. Lisitsyn *, S. V. Tikhonova ** State Research Institute of Aviation Systems, Moscow, Russia * lvm@gosniias.msk.ru

More information

Optical Flow-Based Person Tracking by Multiple Cameras

Optical Flow-Based Person Tracking by Multiple Cameras Proc. IEEE Int. Conf. on Multisensor Fusion and Integration in Intelligent Systems, Baden-Baden, Germany, Aug. 2001. Optical Flow-Based Person Tracking by Multiple Cameras Hideki Tsutsui, Jun Miura, and

More information

P recise Eye. High resolution, diffraction-limited f/4.5 optical quality for high precision measurement and inspection.

P recise Eye. High resolution, diffraction-limited f/4.5 optical quality for high precision measurement and inspection. High resolution, diffraction-limited f/4.5 optical quality for high precision measurement and inspection. Long working distance makes lighting and handling easier. Compact size. Coaxial lighting available

More information

Advanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation

Advanced Vision Guided Robotics. David Bruce Engineering Manager FANUC America Corporation Advanced Vision Guided Robotics David Bruce Engineering Manager FANUC America Corporation Traditional Vision vs. Vision based Robot Guidance Traditional Machine Vision Determine if a product passes or

More information

Recall: Derivative of Gaussian Filter. Lecture 7: Correspondence Matching. Observe and Generalize. Observe and Generalize. Observe and Generalize

Recall: Derivative of Gaussian Filter. Lecture 7: Correspondence Matching. Observe and Generalize. Observe and Generalize. Observe and Generalize Recall: Derivative of Gaussian Filter G x I x =di(x,y)/dx Lecture 7: Correspondence Matching Reading: T&V Section 7.2 I(x,y) G y convolve convolve I y =di(x,y)/dy Observe and Generalize Derivative of Gaussian

More information

Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement

Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement Ta Yuan and Murali Subbarao tyuan@sbee.sunysb.edu and murali@sbee.sunysb.edu Department of

More information

SL A Tordivel - Thor Vollset -Stereo Vision and structured illumination creates dense 3D Images Page 1

SL A Tordivel - Thor Vollset -Stereo Vision and structured illumination creates dense 3D Images Page 1 Tordivel ASTORDIVEL 2000-2015 Scorpion Vision Software Scorpion Stinger are trademarks SL-2010-0001-A AS - Scorpion Visionand 8 and 3DMaMa Tordivel ASof Tordivel AS 2000-2010 Page 1 Stereo Vision and structured

More information

Accelerating Pattern Matching or HowMuchCanYouSlide?

Accelerating Pattern Matching or HowMuchCanYouSlide? Accelerating Pattern Matching or HowMuchCanYouSlide? Ofir Pele and Michael Werman School of Computer Science and Engineering The Hebrew University of Jerusalem {ofirpele,werman}@cs.huji.ac.il Abstract.

More information

Effects Of Shadow On Canny Edge Detection through a camera

Effects Of Shadow On Canny Edge Detection through a camera 1523 Effects Of Shadow On Canny Edge Detection through a camera Srajit Mehrotra Shadow causes errors in computer vision as it is difficult to detect objects that are under the influence of shadows. Shadow

More information

A COMPREHENSIVE SIMULATION SOFTWARE FOR TEACHING CAMERA CALIBRATION

A COMPREHENSIVE SIMULATION SOFTWARE FOR TEACHING CAMERA CALIBRATION XIX IMEKO World Congress Fundamental and Applied Metrology September 6 11, 2009, Lisbon, Portugal A COMPREHENSIVE SIMULATION SOFTWARE FOR TEACHING CAMERA CALIBRATION David Samper 1, Jorge Santolaria 1,

More information

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW Thorsten Thormählen, Hellward Broszio, Ingolf Wassermann thormae@tnt.uni-hannover.de University of Hannover, Information Technology Laboratory,

More information

Creating a distortion characterisation dataset for visual band cameras using fiducial markers.

Creating a distortion characterisation dataset for visual band cameras using fiducial markers. Creating a distortion characterisation dataset for visual band cameras using fiducial markers. Robert Jermy Council for Scientific and Industrial Research Email: rjermy@csir.co.za Jason de Villiers Council

More information

Full Body Tracking Using an Agent-based Architecture

Full Body Tracking Using an Agent-based Architecture Full Body Tracking Using an Agent-based Architecture Bing Fang, Liguang Xie, Pak-Kiu Chung, Yong Cao, Francis Quek Center for Human Computer Interaction Virginia Polytechnic Institute and State University

More information

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016 edestrian Detection Using Correlated Lidar and Image Data EECS442 Final roject Fall 2016 Samuel Rohrer University of Michigan rohrer@umich.edu Ian Lin University of Michigan tiannis@umich.edu Abstract

More information

1 Clustering with K-means

1 Clustering with K-means 1 Clustering with K-means Consider the two datasets fig..a,b and fig..c,d below. We want to cluster these datasets using K-means (L- norm, K = ), with different initial conditions. Starting from the given

More information

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview Human Body Recognition and Tracking: How the Kinect Works Kinect RGB-D Camera Microsoft Kinect (Nov. 2010) Color video camera + laser-projected IR dot pattern + IR camera $120 (April 2012) Kinect 1.5 due

More information

DRC A Multi-Camera System on PC-Cluster for Real-time 3-D Tracking. Viboon Sangveraphunsiri*, Kritsana Uttamang, and Pongsakon Pedpunsri

DRC A Multi-Camera System on PC-Cluster for Real-time 3-D Tracking. Viboon Sangveraphunsiri*, Kritsana Uttamang, and Pongsakon Pedpunsri The 23 rd Conference of the Mechanical Engineering Network of Thailand November 4 7, 2009, Chiang Mai A Multi-Camera System on PC-Cluster for Real-time 3-D Tracking Viboon Sangveraphunsiri*, Kritsana Uttamang,

More information

Carmen Alonso Montes 23rd-27th November 2015

Carmen Alonso Montes 23rd-27th November 2015 Practical Computer Vision: Theory & Applications 23rd-27th November 2015 Wrap up Today, we are here 2 Learned concepts Hough Transform Distance mapping Watershed Active contours 3 Contents Wrap up Object

More information

Improving the 3D Scan Precision of Laser Triangulation

Improving the 3D Scan Precision of Laser Triangulation Improving the 3D Scan Precision of Laser Triangulation The Principle of Laser Triangulation Triangulation Geometry Example Z Y X Image of Target Object Sensor Image of Laser Line 3D Laser Triangulation

More information

Linescan System Design for Robust Web Inspection

Linescan System Design for Robust Web Inspection Linescan System Design for Robust Web Inspection Vision Systems Design Webinar, December 2011 Engineered Excellence 1 Introduction to PVI Systems Automated Test & Measurement Equipment PC and Real-Time

More information

Landmarks for absolute localization

Landmarks for absolute localization Landmarks for absolute localization Jon Howell Keith Kotay Technical Report TR2000-364 Department of Computer Science Dartmouth College Hanover, NH 03755-3510 jonh@cs.dartmouth.edu Abstract For certain

More information

An Image-Based Three-Dimensional Digitizer for Pre-Decorating Thermoformed Parts

An Image-Based Three-Dimensional Digitizer for Pre-Decorating Thermoformed Parts An Image-Based Three-Dimensional Digitizer for Pre-Decorating Thermoformed Parts J.P. Mellor Rose-Hulman Institute of Technology jpmellor@rose-hulman.edu Abstract Thermoformed plastic parts are pervasive

More information

Using Web Camera Technology to Monitor Steel Construction

Using Web Camera Technology to Monitor Steel Construction Using Web Camera Technology to Monitor Steel Construction Kerry T. Slattery, Ph.D., P.E. Southern Illinois University Edwardsville Edwardsville, Illinois Many construction companies install electronic

More information

lecture 10 - depth from blur, binocular stereo

lecture 10 - depth from blur, binocular stereo This lecture carries forward some of the topics from early in the course, namely defocus blur and binocular disparity. The main emphasis here will be on the information these cues carry about depth, rather

More information

Peak Detector. Minimum Detectable Z Step. Dr. Josep Forest Technical Director. Copyright AQSENSE, S.L.

Peak Detector. Minimum Detectable Z Step. Dr. Josep Forest Technical Director. Copyright AQSENSE, S.L. Peak Detector Minimum Detectable Z Step Dr. Josep Forest Technical Director Peak Detector Minimum Detectable Defect Table of Contents 1.Introduction...4 2.Layout...4 3.Results...8 4.Conclusions...9 Copyright

More information

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors

Complex Sensors: Cameras, Visual Sensing. The Robotics Primer (Ch. 9) ECE 497: Introduction to Mobile Robotics -Visual Sensors Complex Sensors: Cameras, Visual Sensing The Robotics Primer (Ch. 9) Bring your laptop and robot everyday DO NOT unplug the network cables from the desktop computers or the walls Tuesday s Quiz is on Visual

More information

Head Pose Estimation by using Morphological Property of Disparity Map

Head Pose Estimation by using Morphological Property of Disparity Map Head Pose Estimation by using Morphological Property of Disparity Map Sewoong Jun*, Sung-Kee Park* and Moonkey Lee** *Intelligent Robotics Research Center, Korea Institute Science and Technology, Seoul,

More information

Bioimage Informatics. Lecture 8, Spring Bioimage Data Analysis (II): Point Feature Detection

Bioimage Informatics. Lecture 8, Spring Bioimage Data Analysis (II): Point Feature Detection Bioimage Informatics Lecture 8, Spring 2012 Bioimage Data Analysis (II): Point Feature Detection Lecture 8 February 13, 2012 1 Outline Project assignment 02 Comments on reading assignment 01 Review: pixel

More information

Perfect Presentations Hop-around Cards

Perfect Presentations Hop-around Cards Perfect Presentations Hop-around Cards Visit us online at HOP-AROUND CARDS Instructions for use. Preparation Print the cards out using a high quality colour printer Laminate each sheet and then cut out

More information

A Study on the Distortion Correction Methodology of Vision Sensor

A Study on the Distortion Correction Methodology of Vision Sensor , July 2-4, 2014, London, U.K. A Study on the Distortion Correction Methodology of Vision Sensor Younghoon Kho, Yongjin (James) Kwon 1 Abstract This study investigates a simple and effective vision calibration

More information

Detecting Fingertip Method and Gesture Usability Research for Smart TV. Won-Jong Yoon and Jun-dong Cho

Detecting Fingertip Method and Gesture Usability Research for Smart TV. Won-Jong Yoon and Jun-dong Cho Detecting Fingertip Method and Gesture Usability Research for Smart TV Won-Jong Yoon and Jun-dong Cho Outline Previous work Scenario Finger detection Block diagram Result image Performance Usability test

More information

CSE152 Introduction to Computer Vision Assignment 3 (SP15) Instructor: Ben Ochoa Maximum Points : 85 Deadline : 11:59 p.m., Friday, 29-May-2015

CSE152 Introduction to Computer Vision Assignment 3 (SP15) Instructor: Ben Ochoa Maximum Points : 85 Deadline : 11:59 p.m., Friday, 29-May-2015 Instructions: CSE15 Introduction to Computer Vision Assignment 3 (SP15) Instructor: Ben Ochoa Maximum Points : 85 Deadline : 11:59 p.m., Friday, 9-May-015 This assignment should be solved, and written

More information

CS4758: Rovio Augmented Vision Mapping Project

CS4758: Rovio Augmented Vision Mapping Project CS4758: Rovio Augmented Vision Mapping Project Sam Fladung, James Mwaura Abstract The goal of this project is to use the Rovio to create a 2D map of its environment using a camera and a fixed laser pointer

More information

Parallel algorithms for stereo vision shape from stereo

Parallel algorithms for stereo vision shape from stereo Parallel algorithms for stereo vision shape from stereo XINGJIE RUAN, XENOPHON Institute of Information & Mathematical Sciences Massey University at Albany, Auckland, New Zealand Xingjie.Ruan.1@uni.massey.ac.nz

More information

CSE152a Computer Vision Assignment 2 WI14 Instructor: Prof. David Kriegman. Revision 1

CSE152a Computer Vision Assignment 2 WI14 Instructor: Prof. David Kriegman. Revision 1 CSE152a Computer Vision Assignment 2 WI14 Instructor: Prof. David Kriegman. Revision 1 Instructions: This assignment should be solved, and written up in groups of 2. Work alone only if you can not find

More information

Robert Collins CSE486, Penn State. Lecture 09: Stereo Algorithms

Robert Collins CSE486, Penn State. Lecture 09: Stereo Algorithms Lecture 09: Stereo Algorithms left camera located at (0,0,0) Recall: Simple Stereo System Y y Image coords of point (X,Y,Z) Left Camera: x T x z (, ) y Z (, ) x (X,Y,Z) z X right camera located at (T x,0,0)

More information

OBSTACLE DETECTION USING STRUCTURED BACKGROUND

OBSTACLE DETECTION USING STRUCTURED BACKGROUND OBSTACLE DETECTION USING STRUCTURED BACKGROUND Ghaida Al Zeer, Adnan Abou Nabout and Bernd Tibken Chair of Automatic Control, Faculty of Electrical, Information and Media Engineering University of Wuppertal,

More information

A Vision System for Automatic State Determination of Grid Based Board Games

A Vision System for Automatic State Determination of Grid Based Board Games A Vision System for Automatic State Determination of Grid Based Board Games Michael Bryson Computer Science and Engineering, University of South Carolina, 29208 Abstract. Numerous programs have been written

More information

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc.

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc. Minimizing Noise and Bias in 3D DIC Correlated Solutions, Inc. Overview Overview of Noise and Bias Digital Image Correlation Background/Tracking Function Minimizing Noise Focus Contrast/Lighting Glare

More information

CS5670: Computer Vision

CS5670: Computer Vision CS5670: Computer Vision Noah Snavely Light & Perception Announcements Quiz on Tuesday Project 3 code due Monday, April 17, by 11:59pm artifact due Wednesday, April 19, by 11:59pm Can we determine shape

More information

NAME VCamera camera model representation

NAME VCamera camera model representation NAME VCamera camera model representation SYNOPSIS #include void VRegisterCameraType (void); extern VRepnKind VCameraRepn; VCamera camera; ld... -lvcam -lvista -lm -lgcc... DESCRIPTION

More information

Horus: Object Orientation and Id without Additional Markers

Horus: Object Orientation and Id without Additional Markers Computer Science Department of The University of Auckland CITR at Tamaki Campus (http://www.citr.auckland.ac.nz) CITR-TR-74 November 2000 Horus: Object Orientation and Id without Additional Markers Jacky

More information

Towards the completion of assignment 1

Towards the completion of assignment 1 Towards the completion of assignment 1 What to do for calibration What to do for point matching What to do for tracking What to do for GUI COMPSCI 773 Feature Point Detection Why study feature point detection?

More information

Lumascope Startup Guide -- Lumascope 720

Lumascope Startup Guide -- Lumascope 720 Lumascope 720 -- Startup Guide -- Lumascope 720 IMPORTANT: This Guide describes how to set up the LS720 (Section I) and start using Lumaview 720-Series (Section II). It is important that you follow this

More information

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: , 3D Sensing and Reconstruction Readings: Ch 12: 12.5-6, Ch 13: 13.1-3, 13.9.4 Perspective Geometry Camera Model Stereo Triangulation 3D Reconstruction by Space Carving 3D Shape from X means getting 3D coordinates

More information

Gregory Walsh, Ph.D. San Ramon, CA January 25, 2011

Gregory Walsh, Ph.D. San Ramon, CA January 25, 2011 Leica ScanStation:: Calibration and QA Gregory Walsh, Ph.D. San Ramon, CA January 25, 2011 1. Summary Leica Geosystems, in creating the Leica Scanstation family of products, has designed and conducted

More information

Development of Vision System on Humanoid Robot HRP-2

Development of Vision System on Humanoid Robot HRP-2 Development of Vision System on Humanoid Robot HRP-2 Yutaro Fukase Institute of Technology, Shimizu Corporation, Japan fukase@shimz.co.jp Junichiro Maeda Institute of Technology, Shimizu Corporation, Japan

More information

COMPUTER-BASED WORKPIECE DETECTION ON CNC MILLING MACHINE TOOLS USING OPTICAL CAMERA AND NEURAL NETWORKS

COMPUTER-BASED WORKPIECE DETECTION ON CNC MILLING MACHINE TOOLS USING OPTICAL CAMERA AND NEURAL NETWORKS Advances in Production Engineering & Management 5 (2010) 1, 59-68 ISSN 1854-6250 Scientific paper COMPUTER-BASED WORKPIECE DETECTION ON CNC MILLING MACHINE TOOLS USING OPTICAL CAMERA AND NEURAL NETWORKS

More information

Introduction to 3D Machine Vision

Introduction to 3D Machine Vision Introduction to 3D Machine Vision 1 Many methods for 3D machine vision Use Triangulation (Geometry) to Determine the Depth of an Object By Different Methods: Single Line Laser Scan Stereo Triangulation

More information

AN AUTOMATIC 3D RECONSTRUCTION METHOD BASED ON MULTI-VIEW STEREO VISION FOR THE MOGAO GROTTOES

AN AUTOMATIC 3D RECONSTRUCTION METHOD BASED ON MULTI-VIEW STEREO VISION FOR THE MOGAO GROTTOES The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume XL-4/W5, 05 Indoor-Outdoor Seamless Modelling, Mapping and avigation, May 05, Tokyo, Japan A AUTOMATIC

More information

International Journal of Advance Engineering and Research Development

International Journal of Advance Engineering and Research Development Scientific Journal of Impact Factor (SJIF): 4.14 International Journal of Advance Engineering and Research Development Volume 3, Issue 3, March -2016 e-issn (O): 2348-4470 p-issn (P): 2348-6406 Research

More information

Work with Shapes. Concepts CHAPTER. Concepts, page 3-1 Procedures, page 3-5

Work with Shapes. Concepts CHAPTER. Concepts, page 3-1 Procedures, page 3-5 3 CHAPTER Revised: November 15, 2011 Concepts, page 3-1, page 3-5 Concepts The Shapes Tool is Versatile, page 3-2 Guidelines for Shapes, page 3-2 Visual Density Transparent, Translucent, or Opaque?, page

More information

Ariel Dynamics, Inc. TRIM MODULE. Revision 1.0. Ariel Dynamics, Inc. C3D TRANSFORM MODULE

Ariel Dynamics, Inc. TRIM MODULE. Revision 1.0. Ariel Dynamics, Inc. C3D TRANSFORM MODULE Ariel Dynamics, Inc. TRIM MODULE Revision 1.0 Ariel Dynamics, Inc. C3D TRANSFORM MODULE Contents i Contents ARIEL TRIM PROGRAM 1 INTRODUCTION...1 WHAT S NEW IN TRIM 1.0...1 SYSTEM REQUIREMENTS...2 TO START

More information

Tracking Under Low-light Conditions Using Background Subtraction

Tracking Under Low-light Conditions Using Background Subtraction Tracking Under Low-light Conditions Using Background Subtraction Matthew Bennink Clemson University Clemson, South Carolina Abstract A low-light tracking system was developed using background subtraction.

More information

Practice Exam Sample Solutions

Practice Exam Sample Solutions CS 675 Computer Vision Instructor: Marc Pomplun Practice Exam Sample Solutions Note that in the actual exam, no calculators, no books, and no notes allowed. Question 1: out of points Question 2: out of

More information

Skeleton Cube for Lighting Environment Estimation

Skeleton Cube for Lighting Environment Estimation (MIRU2004) 2004 7 606 8501 E-mail: {takesi-t,maki,tm}@vision.kuee.kyoto-u.ac.jp 1) 2) Skeleton Cube for Lighting Environment Estimation Takeshi TAKAI, Atsuto MAKI, and Takashi MATSUYAMA Graduate School

More information

Measurement and Precision Analysis of Exterior Orientation Element Based on Landmark Point Auxiliary Orientation

Measurement and Precision Analysis of Exterior Orientation Element Based on Landmark Point Auxiliary Orientation 2016 rd International Conference on Engineering Technology and Application (ICETA 2016) ISBN: 978-1-60595-8-0 Measurement and Precision Analysis of Exterior Orientation Element Based on Landmark Point

More information

AUTOMATIC MODELLING METHOD FOR STEEL STRUCTURES USING PHOTOGRAMMETRY

AUTOMATIC MODELLING METHOD FOR STEEL STRUCTURES USING PHOTOGRAMMETRY AUTOMATIC MODELLING METHOD FOR STEEL STRUCTURES USING PHOTOGRAMMETRY S. H. Jung a, J. H. Yu a*, J. K. Lee b, L. Ge a a School of Surveying and Spatial Information Systems, University of New South Wales,

More information

COMPARATIVE STUDY OF IMAGE EDGE DETECTION ALGORITHMS

COMPARATIVE STUDY OF IMAGE EDGE DETECTION ALGORITHMS COMPARATIVE STUDY OF IMAGE EDGE DETECTION ALGORITHMS Shubham Saini 1, Bhavesh Kasliwal 2, Shraey Bhatia 3 1 Student, School of Computing Science and Engineering, Vellore Institute of Technology, India,

More information

Tracking Trajectories of Migrating Birds Around a Skyscraper

Tracking Trajectories of Migrating Birds Around a Skyscraper Tracking Trajectories of Migrating Birds Around a Skyscraper Brian Crombie Matt Zivney Project Advisors Dr. Huggins Dr. Stewart Abstract In this project, the trajectories of birds are tracked around tall

More information

A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect

A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect A Comparison between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect Diana Beltran and Luis Basañez Technical University of Catalonia, Barcelona, Spain {diana.beltran,luis.basanez}@upc.edu

More information

BIN PICKING APPLICATIONS AND TECHNOLOGIES

BIN PICKING APPLICATIONS AND TECHNOLOGIES BIN PICKING APPLICATIONS AND TECHNOLOGIES TABLE OF CONTENTS INTRODUCTION... 3 TYPES OF MATERIAL HANDLING... 3 WHOLE BIN PICKING PROCESS... 4 VISION SYSTEM: HARDWARE... 4 VISION SYSTEM: SOFTWARE... 5 END

More information

Discovering Visual Hierarchy through Unsupervised Learning Haider Razvi

Discovering Visual Hierarchy through Unsupervised Learning Haider Razvi Discovering Visual Hierarchy through Unsupervised Learning Haider Razvi hrazvi@stanford.edu 1 Introduction: We present a method for discovering visual hierarchy in a set of images. Automatically grouping

More information

Mobile Camera Based Calculator

Mobile Camera Based Calculator Mobile Camera Based Calculator Liwei Wang Jingyi Dai Li Du Department of Electrical Engineering Department of Electrical Engineering Department of Electrical Engineering Stanford University Stanford University

More information

MOVING OBJECT DETECTION USING BACKGROUND SUBTRACTION ALGORITHM USING SIMULINK

MOVING OBJECT DETECTION USING BACKGROUND SUBTRACTION ALGORITHM USING SIMULINK MOVING OBJECT DETECTION USING BACKGROUND SUBTRACTION ALGORITHM USING SIMULINK Mahamuni P. D 1, R. P. Patil 2, H.S. Thakar 3 1 PG Student, E & TC Department, SKNCOE, Vadgaon Bk, Pune, India 2 Asst. Professor,

More information

Miniaturized Camera Systems for Microfactories

Miniaturized Camera Systems for Microfactories Miniaturized Camera Systems for Microfactories Timo Prusi, Petri Rokka, and Reijo Tuokko Tampere University of Technology, Department of Production Engineering, Korkeakoulunkatu 6, 33720 Tampere, Finland

More information

Using Layered Color Precision for a Self-Calibrating Vision System

Using Layered Color Precision for a Self-Calibrating Vision System ROBOCUP2004 SYMPOSIUM, Instituto Superior Técnico, Lisboa, Portugal, July 4-5, 2004. Using Layered Color Precision for a Self-Calibrating Vision System Matthias Jüngel Institut für Informatik, LFG Künstliche

More information

Project 1 : Dead Reckoning and Tracking

Project 1 : Dead Reckoning and Tracking CS3630 Spring 2012 Project 1 : Dead Reckoning and Tracking Group : Wayward Sons Sameer Ansari, David Bernal, Tommy Kazenstein 2/8/2012 Wayward Sons CS3630 Spring 12 Project 1 Page 2 of 12 CS 3630 (Spring

More information

Advances in Metrology for Guide Plate Analysis

Advances in Metrology for Guide Plate Analysis Advances in Metrology for Guide Plate Analysis Oxford Lasers Ltd Overview Context and motivation Latest advances: Automatic entrance hole measurement Hole shape analysis Debris detection File format We

More information

Cognex Machine Vision & Mitsubishi Robot Interface/Setup Tutorial Nicholas Bozzelli & Quentin Kilgore Millersville University October 2017

Cognex Machine Vision & Mitsubishi Robot Interface/Setup Tutorial Nicholas Bozzelli & Quentin Kilgore Millersville University October 2017 Cognex Machine Vision & Mitsubishi Robot Interface/Setup Tutorial Nicholas Bozzelli & Quentin Kilgore Millersville University October 2017 Part I: Cognex Set-Up Overview: Using the Cognex vision system,

More information

Texture Generation for the Computer Representation of the Upper Gastrointestinal System

Texture Generation for the Computer Representation of the Upper Gastrointestinal System A. Gastélum, P. Delmas, J. Márquez, Texture Generation for the Computer Representation of the Upper Gastrointestinal System, Proceedings of Image and Vision Computing New Zealand 2007, pp. 305 310, Hamilton,

More information

The Maze Runner Zumo version. Alexander Kirillov

The Maze Runner Zumo version. Alexander Kirillov The Maze Runner Zumo version Alexander Kirillov URL: http://sigmacamp.org/mazerunner-zumo E-mail address: shurik179@gmail.com This work is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike

More information

Digital Volume Correlation for Materials Characterization

Digital Volume Correlation for Materials Characterization 19 th World Conference on Non-Destructive Testing 2016 Digital Volume Correlation for Materials Characterization Enrico QUINTANA, Phillip REU, Edward JIMENEZ, Kyle THOMPSON, Sharlotte KRAMER Sandia National

More information

Contents. Introduction.1. Specification Data and Package.2. Working Environment Software Installation 3. Hardware Installation.

Contents. Introduction.1. Specification Data and Package.2. Working Environment Software Installation 3. Hardware Installation. User Manual V 1.0N Contents Introduction.1 Specification Data and Package.2 Working Environment...... 3 Software Installation 3 Hardware Installation.4 Calibration Software Manual 7 tboard Software Manual..

More information

Fast and Robust Stereo Object Recognition for Spheres

Fast and Robust Stereo Object Recognition for Spheres Fast and Robust Stereo Object Recognition for Spheres Robby McKilliam and Gordon Wyeth Information Technology and Electrical Engineering University of Queensland, Australia wyeth@itee.uq.edu.au Abstract

More information

Multiple View Reconstruction of Calibrated Images using Singular Value Decomposition

Multiple View Reconstruction of Calibrated Images using Singular Value Decomposition Multiple View Reconstruction of Calibrated Images using Singular Value Decomposition Ayan Chaudhury, Abhishek Gupta, Sumita Manna, Subhadeep Mukherjee, Amlan Chakrabarti Abstract Calibration in a multi

More information

Technical Information. Command overview of Vision Systems

Technical Information. Command overview of Vision Systems Technical Information Command overview of Vision Systems Image analysis command Grab image 0x01 X X X X Shutter speed 0x07 X X X X Synchronous flash 0x49 X X X X Video mode 0x00 X X Display 0x05 X X X

More information

A Real Time Vision System for Robotic Soccer

A Real Time Vision System for Robotic Soccer A Real Time Vision System for Robotic Soccer Chunmiao Wang, Hui Wang, William Y. C. Soh, Han Wang Division of Control & Instrumentation, School of EEE, Nanyang Technological University, 4 Nanyang Avenue,

More information

Final Exam Study Guide

Final Exam Study Guide Final Exam Study Guide Exam Window: 28th April, 12:00am EST to 30th April, 11:59pm EST Description As indicated in class the goal of the exam is to encourage you to review the material from the course.

More information

A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision

A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision A Simple Interface for Mobile Robot Equipped with Single Camera using Motion Stereo Vision Stephen Karungaru, Atsushi Ishitani, Takuya Shiraishi, and Minoru Fukumi Abstract Recently, robot technology has

More information

Subject-Oriented Image Classification based on Face Detection and Recognition

Subject-Oriented Image Classification based on Face Detection and Recognition 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

6.2 Classification of Closed Surfaces

6.2 Classification of Closed Surfaces Table 6.1: A polygon diagram 6.1.2 Second Proof: Compactifying Teichmuller Space 6.2 Classification of Closed Surfaces We saw that each surface has a triangulation. Compact surfaces have finite triangulations.

More information

Calibration of Video Cameras to the Coordinate System of a Radiation Therapy Treatment Machine

Calibration of Video Cameras to the Coordinate System of a Radiation Therapy Treatment Machine Calibration of Video Cameras to the Coordinate System of a Radiation Therapy Treatment Machine Scott W. Hadley, L. Scott Johnson, and Charles A. Pelizzari University of Chicago The Department of Radiation

More information

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation. 3D Sensing 3D Shape from X Perspective Geometry Camera Model Camera Calibration General Stereo Triangulation 3D Reconstruction 3D Shape from X shading silhouette texture stereo light striping motion mainly

More information

Digital Image Correlation Compared to Strain Gauge

Digital Image Correlation Compared to Strain Gauge Digital Image Correlation Compared to Strain Gauge Abstract: This report presents the test results from a stereo digital image correlation system, in which two samples were pulled on a tensile test machine.

More information

The Maze Runner. Alexander Kirillov

The Maze Runner. Alexander Kirillov The Maze Runner URL: http://sigmacamp.org/mazerunner E-mail address: shurik179@gmail.com Alexander Kirillov This work is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License.

More information

Top joint. Middle joint. Bottom joint. (joint angle (2)) Finger axis. (joint angle (1)) F (finger frame) P (palm frame) R, t (palm pose) Camera frame

Top joint. Middle joint. Bottom joint. (joint angle (2)) Finger axis. (joint angle (1)) F (finger frame) P (palm frame) R, t (palm pose) Camera frame Hand Pose Recovery with a Single Video Camera Kihwan Kwon, Hong Zhang, and Fadi Dornaika Dept. of Computing Science, University of Alberta Edmonton, Alberta, Canada fkihwan, zhang, dornaikag@cs.ualberta.ca

More information

Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System

Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System Proc. of IEEE Conference on Computer Vision and Pattern Recognition, vol.2, II-131 II-137, Dec. 2001. Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System

More information

Discuss Proven technologies that addresses

Discuss Proven technologies that addresses Robotics and Machine Vision for assembly -Auto Teach, Vision guidance, Color & 3D Mar 5-12 2007 Agenda Discuss Proven technologies that addresses o Quick Tool Bring up o Using Non-touch Vision based Auto

More information