A multi-camera positioning system for steering of a THz stand-off scanner

Size: px
Start display at page:

Download "A multi-camera positioning system for steering of a THz stand-off scanner"

Transcription

1 A multi-camera positioning system for steering of a THz stand-off scanner Maria Axelsson, Mikael Karlsson and Staffan Rudner Swedish Defence Research Agency, Box 1165, SE Linköping, SWEDEN ABSTRACT Stand-off THz imaging to detect concealed treats is a coming technique for security applications. A THz sensor can provide high resolution 3D imagery of a scene. However, efficient scene scanning and management of the THz sensor is a challenging task due to the limited field of view of the sensor and physical scanning limitations. In this paper we discuss the requirements on a scene scanning solution and present a scene scanning technique using a multi-camera system with 3D positioning capabilities. A visual hull method is used to position subjects in the scene. The presented technique relaxes the requirements on the scanning speed of the THz sensor and facilitates an efficient scene scanning solution. Keywords: Imaging, screening, stand-off detection, teraherz, visual hull 1. INTRODUCTION Detection of concealed threats at stand-off distance is desired in many security applications. Upcoming techniques for stand-off detection use sub-millimeter-wave imaging systems which can provide high resolution 3D imagery of a scene. In such 3D imagery hidden threats can be detected using manual or automated methods and security can be alerted before the threat is close. The THz sensor systems developed for stand-off detection usually has a limited spatial coverage due to the instantaneous field of view, the scanning rate, and the manageable data rate, which puts a demand on methods for efficient scene scanning. A scene scanning system is needed to control the THz sensor and steer it to the 3D position where a subject or specific part of a subject is located. Valuable imaging time can be saved using accurate 3D positioning of each subject to avoid scanning empty areas. Subjects can be tracked and positioned in the scene and at each time the sensor can be steered to the next point of interest. In complex scenarios where subjects are allowed to walk or move freely, the system can also be used to track individual body parts and obtain full coverage of the body incrementally, as different body parts of a person become visible to the THz sensor system. In addition to positioning, a scene scanning system can also provide an estimate of scan completeness or information from the scanning system can be used to merge the high resolution volume data from different scans, e.g., frontal and rear scans of a person, to verify that full coverage of the body is obtained. In this paper we discuss scene scanning methods using multi-sensor approaches. We also present our experimental results from an investigation of a scene scanning technique using a multi-camera system with 3D positioning capabilities. Multi-sensor positioning systems are investigated since their ability to detect and position subjects accurately in 3D is greatly improved compared to using only a single sensor. The 3D positioning is demonstrated on real data acquired from seven HD video cameras. We use a visual hull method where each camera view provide support for the presence of interesting foreground objects in each part of the scene based on an adaptive Gaussian background model. The presented technique relaxes the requirements on the scanning speed of the THz sensor and facilitates an efficient scene scanning solution. Correspondence: maria.axelsson@foi.se Passive Millimeter-Wave Imaging Technology XIV, edited by David A. Wikner, Arttu R. Luukanen, Proc. of SPIE Vol. 8022, 80220L 2011 SPIE CCC code: X/11/$18 doi: / Proc. of SPIE Vol L-1

2 Figure 1. Example scene where a single subject is scanned from three sides while walking through an appointed path. 2. SCANNING SCENARIOS Several scenarios can be imagined at a security checkpoint. The problem complexity increases with the number of subjects in the scene and with their allowed variation in pose and position. The simplest case is where a single subject stands in an appointed pose while the THz system scans one side. The subject then turns and the other side is scanned. As the pose and position is roughly known the scene scanning system has much a-priori information. If the high resolution sensor is slow in image acquisition compared to the time it is possible to stand still, it becomes necessary to track body parts and keep track of the scanned volumes. The problem complexity increases when the subject is allowed to move more freely. Figure 1 shows an example scene setup for a case where the subject is moving along a path and scans can be obtained from three different views during motion. The position and pose of a subject walking through the security checkpoint need to be tracked to ensure that all parts of the body are scanned. Depending on the time constraints set by the speed of the THz scanning system, both the position and orientation of the person and the positions and orientations of the arms and legs might be needed. In an environment where one person at the time passes through the scanning area, there is no risk of interperson occlusions. Additionally, the scanning system does not need to shift its focus between multiple persons. This relaxes the requirement on fast sensor movements. It is desirable maximize the flow through a security checkpoint, and hence unnecessary restrictions should be avoided. If multiple subjects are allowed to walk along the designated path simultaneously, there is of course an increase in complexity and such scenarios are further restricted by the image acquisition time. In an environment where subjects can move entirely freely, occlusions are likely to preclude complete scanning of all persons. The highest complexity level is scanning a scene with a free flowing crowd. This is far beyond the achievable horizon at this time. Proc. of SPIE Vol L-2

3 3. CAMERA CONFIGURATION FOR A SCENE SCANNING SYSTEM Information about the position and pose of all subjects in a scene can be obtained using many different sensor configurations. We have briefly looked into different sensors, e.g., visual cameras, infrared cameras, and range sensors. However, as many image processing methods are already available for visual cameras we decided to start our experiments with a multi-camera setup using visual cameras. Infrared cameras can also be added to aid detection of subjects. A camera configuration should contain multiple cameras to facilitate positioning of individuals that needs to be screened. When using multiple single cameras, positioning and tracking performance is improved compared to using a single sensor. Both occlusions between subjects and subject self-occlusion can be handled better using a multi-camera setup. If multiple cameras are used, distance measurements can be made using triangulation, stereo images (dense stereo maps) can be calculated using pairs of cameras to identify several individuals which are partly occluded easier, and 3D positioning methods like the visual hull can be used. The method using visual hull is further described in Section 4. The camera positions must be known to be able to calculate stereo maps and triangulate distances to objects in the world. This requires that both the intrinsic and extrinsic camera parameters are known. An example of practical camera calibration is described in Section 5. The cameras should be set up in a configuration around the scene where they cover a common field of view. In our experiment described in Section 5 we have used seven cameras in a circle to be able to investigate methods for pose and positioning using a multiple-view setup. However, in a final scanning system fewer cameras in a half circle where one is positioned close to the high resolution sensor may be enough. This will be determined by demands from the application on the scene scanning and the requirements on the tracking and positioning algorithms. 4. IMAGE PROCESSING METHODS Image processing methods can be used for 3D positioning, shape extraction, and pose estimation using the camera data. In addition, the data can be used to extract 3D models of each subject to be matched with the high resolution THz data and gain information about the completeness of the scans. Some of the available image processing methods are described in the following paragraphs. First detection of individuals in the scene is needed. This can be achieved by using a background model and extract the foreground in each frame. Then potential targets can be detected using a tailored detector, e.g., a head detector. Accurate and consistent detections in complex scenes with many subjects or many moving objects are usually difficult to obtain. Therefore detections are often not used by themselves if a robust positioning method is needed. Instead, detections can be passed on to a target tracker which use a model of the possible movements to select feasible tracks. If the scanning system is fast and can capture the interesting part of a body before the subject has moved too far, this sort of tracking together with a silhouette image from the position of the THz sensor may be enough in a scene scanning system. If more precise 3D positioning is needed, full body volume detection and tracking of body parts can be used. An example method for 3D positioning and estimation of shape is extraction of the visual hull. Also here a background model is used and the foreground objects, such as individuals, are segmented in each view and segmentations from all camera views are used to reconstruct the 3D bounding volume of the foreground objects. This is in some sense similar to tomographic reconstruction except that we only have binary images with object and background classes. The visual hull was introduced by Laurentini 1 as the volume which completely encloses an object in a scene given a set of silhouette images. Is is widely used to produce three dimensional models from multiple views. In multi-person scenarios, many persons often occlude each other in some camera views and therefore this type of method often needs many views, which may be impractical. To relax the requirements on the segmentation of the subjects against the background, range sensors can be used to aid the extraction of the visual hull. In addition to positioning subjects, pose estimation is important if subjects are allowed to move in the scene or if the scanning is very slow and subjects can not stand still. In such scenarios data from high resolution Proc. of SPIE Vol L-3

4 Figure 2. The actors in the field experiment showing the variation of clothes. imagery must be registered to the correct body parts on a model to identify regions that are not yet scanned. Pose estimation can, e.g., be performed by extracting silhouettes from all camera images and matching them to possible poses, selecting the pose which best matches the silhouettes. Pose can also be obtained using dense stereo map or range sensors.2 5. EXPERIMENTAL SETUP A field trial was carried out in the summer of The aim of the trial was to collect data sets to assess scene scanning methods and evaluate specific image processing methods. In the experiment seven Panasonic HDC SD-700 camcorders recording at 50 fps were used. A fictive circular shaped checkpoint with approximately 12 meters in diameter was built. Cameras and lightning were rigged in a circle, where the cameras were mounted meters above the ground facing a common scene center. Measurements were collected during two days. The first day our three main experiments where carried out where four actors walked along three pre-defined paths. The actors wore more and less concealing clothes to get data suitable for evaluation of pose estimation difficulty. Figure 2 shows the actors and three sets of clothes. The second day, a smaller experiment where four different actors in regular office clothes were standing in different simple poses to use for initial experiments with pose tracking. Camera calibration scenes were also recorded both days. First the video sequences were time synchronized. An example frame is shown in Figure 3. This was performed by detecting a specific event in all views such as a clap and flashing light. The frame rate is high which makes accurate time synchronization easier. The intrinsic camera parameters (focal length, principal point, and lens distortion) and the extrinsic camera parameters (camera position and orientation) of the seven cameras must be determined to be able to obtain accurate measurements from the video sequences. The intrinsic parameters are calibrated for each camera individually using the camera calibration toolbox for Matlab.3 A checkerboard pattern is moved around to cover Proc. of SPIE Vol L-4

5 Figure 3. Seven cameras are synchronized and calibrated to obtain measurements of a common scene. the entire field of view of the camera. The corners on the checkerboard are found automatically4 and extracted with sub-pixel precision. A wrapper is used for the main calibration function in the toolbox to automate the intrinsic camera calibration. The extrinsic camera parameters were determined using reference markers on the ground. The coordinates of the markers on the ground, the coordinates of the markers in each camera view and the intrinsic camera parameters are then used together to calculate the position and orientation of the each camera. The current extrinsic calibration method has an accuracy below a couple of centimeters at a distance of five to six meters (in the scene center). The manual steps in the current calibration method can be reduced and the accuracy can be improved. We are currently investigating automated self-calibration of the rig to facilitate fast deployment of the scene scanning system.5 6. RESULTS FROM THE EXPERIMENTS The measurements from the field trial have been used to evaluate a method for scene scanning using extraction of the visual hull of the subjects in the scene. The visual hull is the 3D space which is bounded by the projection of the subject in the camera views. It can be refined when more cameras are added to a scene. The method we use to extract the visual hull is based on silhouette images of the subject. Each of the seven camera views in a frame is segmented into foreground and background using a Gaussian background model. All pixels in the image are given a probability of belonging to the foreground. The background model is also adaptive to gradual changes over time. An example image and the probability of each pixel belonging to the foreground is shown in Figure 4. The visual hull is extracted by projecting the foreground pixels in each camera to the common world coordinates in the scene. The world is divided into cells which gain support from each camera that there is an object of interest occupying the cell. If many (often all) cameras agree that there is something in the cell it is set as object. The visual hull extraction is calculated in levels from the ground and upwards and the resolution of the cells can be set differently for all three dimensions. The resulting visual hull is represented as cells which are set or not set. It is also possible to set a probability for each cell to contain an object. A visualization of the extracted visual hull is shown in Figure 5. As can be seen in the figure shadows can affect the extracted visual hull. Also pixels misclassified as foreground or background effect the result. Proc. of SPIE Vol L-5

6 Figure 4. (Left) Image frame from one of the cameras. (Right) Foreground segmentation using the Gaussian background model. Figure 5. (Left) Visualization of the extracted visual hull and its position on the floor in 3D. (Right) A part of a camera frame used in the reconstruction of the visual hull. These effects can be reduced using an improved model for background extraction. Our example here is used for demonstration purposes and the segmentation parameters have not been tuned to our specific conditions at this time. Further visualization examples of the visual hull are shown in Figure 6. All examples of the visual hull shown here are extracted using a cell size of 0.01 meters. This small cell size may not be required for an efficient scene scanning method. However, this needs further investigation. A method for scene scanning using visual hull may be used in combination with target tracking to find the interesting part of the scene where accurate 3D positioning is needed. This can reduce the processing time required for the scene scanning. The time requirements on the scene scanning system to output new scanning positions will depend on the scanning speed of the THz system. If the scanning is fast the requirement on tracking body parts is somewhat relaxed and if the scanning is slow there will be more time left to process the images and track subjects and their body parts. Proc. of SPIE Vol L-6

7 Figure 6. The extracted visual hull of a subject from four different views. 7. CONCLUSION AND FUTURE WORK In this paper we have discussed some of the available techniques for scene scanning. With the data from the field trial we have demonstrated a method for 3D positioning of subjects in a scene, using a visual hull method. This type of method is feasible for scenes with few well separated individuals since occlusion between subjects may create ghost answers. If the scanning using the THz system is rapid, scenarios like a single subject standing in any pose may be solvable in combination with a merging of data from, e.g., the front and back side of the subject. To handle more complex scenes, e.g., where a single subject moves along an appointed path, the visual hull method needs to be extended with tracking functionality. Other methods such as pose estimation from silhouettes may be possible to use directly without extracting the full visual hull as an intermediate step. Pose estimation can also be used to assess scan completeness. Further investigations are needed to evaluate the methods suggested for scene scanning with respect to performance and speed. ACKNOWLEDGMENTS MSB (Swedish Civil Contingencies Agency) and FMV (Swedish Defence Materiel Administration) are acknowledged for funding this study. REFERENCES 1. A. Laurentini, The visual hull concept for silhouette-based image understanding, IEEE Trans. Pattern Anal. Mach. Intell. 16, pp , February J. Shotton, A. Fitzgibbon, M. Cook, T. Sharp, M. Finocchio, R. Moore, A. Kipman, and A. Blake, Real-time human pose recognition in parts from single depth images, in IEEE International Conference on Computer Vision, J.-Y. Bouguet, Camera calibration toolbox for Matlab. doc/. 4. M. Axelsson, P. Follo, and C. Grönwall, Camera calibration using automated identification of checkerboard patterns, in Proceedings of SSBA 2010, Symposium on image analysis, Uppsala, M. Axelsson, Automatic calibration of a camera positioning system using estimation of the essential matrix, in Proceedings of SSBA 2011, Symposium on image analysis, Linköping, Proc. of SPIE Vol L-7

Lecture 8 Active stereo & Volumetric stereo

Lecture 8 Active stereo & Volumetric stereo Lecture 8 Active stereo & Volumetric stereo Active stereo Structured lighting Depth sensing Volumetric stereo: Space carving Shadow carving Voxel coloring Reading: [Szelisky] Chapter 11 Multi-view stereo

More information

Three-dimensional nondestructive evaluation of cylindrical objects (pipe) using an infrared camera coupled to a 3D scanner

Three-dimensional nondestructive evaluation of cylindrical objects (pipe) using an infrared camera coupled to a 3D scanner Three-dimensional nondestructive evaluation of cylindrical objects (pipe) using an infrared camera coupled to a 3D scanner F. B. Djupkep Dizeu, S. Hesabi, D. Laurendeau, A. Bendada Computer Vision and

More information

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Tomokazu Satoy, Masayuki Kanbaray, Naokazu Yokoyay and Haruo Takemuraz ygraduate School of Information

More information

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera Tomokazu Sato, Masayuki Kanbara and Naokazu Yokoya Graduate School of Information Science, Nara Institute

More information

Project Title: Welding Machine Monitoring System Phase II. Name of PI: Prof. Kenneth K.M. LAM (EIE) Progress / Achievement: (with photos, if any)

Project Title: Welding Machine Monitoring System Phase II. Name of PI: Prof. Kenneth K.M. LAM (EIE) Progress / Achievement: (with photos, if any) Address: Hong Kong Polytechnic University, Phase 8, Hung Hom, Kowloon, Hong Kong. Telephone: (852) 3400 8441 Email: cnerc.steel@polyu.edu.hk Website: https://www.polyu.edu.hk/cnerc-steel/ Project Title:

More information

Visual Hulls from Single Uncalibrated Snapshots Using Two Planar Mirrors

Visual Hulls from Single Uncalibrated Snapshots Using Two Planar Mirrors Visual Hulls from Single Uncalibrated Snapshots Using Two Planar Mirrors Keith Forbes 1 Anthon Voigt 2 Ndimi Bodika 2 1 Digital Image Processing Group 2 Automation and Informatics Group Department of Electrical

More information

Pattern Feature Detection for Camera Calibration Using Circular Sample

Pattern Feature Detection for Camera Calibration Using Circular Sample Pattern Feature Detection for Camera Calibration Using Circular Sample Dong-Won Shin and Yo-Sung Ho (&) Gwangju Institute of Science and Technology (GIST), 13 Cheomdan-gwagiro, Buk-gu, Gwangju 500-71,

More information

Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement

Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement Integration of Multiple-baseline Color Stereo Vision with Focus and Defocus Analysis for 3D Shape Measurement Ta Yuan and Murali Subbarao tyuan@sbee.sunysb.edu and murali@sbee.sunysb.edu Department of

More information

Measurement of 3D Foot Shape Deformation in Motion

Measurement of 3D Foot Shape Deformation in Motion Measurement of 3D Foot Shape Deformation in Motion Makoto Kimura Masaaki Mochimaru Takeo Kanade Digital Human Research Center National Institute of Advanced Industrial Science and Technology, Japan The

More information

Creating a distortion characterisation dataset for visual band cameras using fiducial markers.

Creating a distortion characterisation dataset for visual band cameras using fiducial markers. Creating a distortion characterisation dataset for visual band cameras using fiducial markers. Robert Jermy Council for Scientific and Industrial Research Email: rjermy@csir.co.za Jason de Villiers Council

More information

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation.

3D Sensing. 3D Shape from X. Perspective Geometry. Camera Model. Camera Calibration. General Stereo Triangulation. 3D Sensing 3D Shape from X Perspective Geometry Camera Model Camera Calibration General Stereo Triangulation 3D Reconstruction 3D Shape from X shading silhouette texture stereo light striping motion mainly

More information

Walking gait dataset: point clouds, skeletons and silhouettes

Walking gait dataset: point clouds, skeletons and silhouettes Walking gait dataset: point clouds, skeletons and silhouettes Technical Report Number 1379 Trong-Nguyen Nguyen * and Jean Meunier DIRO, University of Montreal, Montreal, QC, Canada September 8, 2018 Abstract

More information

A 3-D Scanner Capturing Range and Color for the Robotics Applications

A 3-D Scanner Capturing Range and Color for the Robotics Applications J.Haverinen & J.Röning, A 3-D Scanner Capturing Range and Color for the Robotics Applications, 24th Workshop of the AAPR - Applications of 3D-Imaging and Graph-based Modeling, May 25-26, Villach, Carinthia,

More information

Fundamental Matrices from Moving Objects Using Line Motion Barcodes

Fundamental Matrices from Moving Objects Using Line Motion Barcodes Fundamental Matrices from Moving Objects Using Line Motion Barcodes Yoni Kasten (B), Gil Ben-Artzi, Shmuel Peleg, and Michael Werman School of Computer Science and Engineering, The Hebrew University of

More information

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview

Human Body Recognition and Tracking: How the Kinect Works. Kinect RGB-D Camera. What the Kinect Does. How Kinect Works: Overview Human Body Recognition and Tracking: How the Kinect Works Kinect RGB-D Camera Microsoft Kinect (Nov. 2010) Color video camera + laser-projected IR dot pattern + IR camera $120 (April 2012) Kinect 1.5 due

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Surround Structured Lighting for Full Object Scanning

Surround Structured Lighting for Full Object Scanning Surround Structured Lighting for Full Object Scanning Douglas Lanman, Daniel Crispell, and Gabriel Taubin Department of Engineering, Brown University {dlanman,daniel crispell,taubin}@brown.edu Abstract

More information

Surround Structured Lighting for Full Object Scanning

Surround Structured Lighting for Full Object Scanning Surround Structured Lighting for Full Object Scanning Douglas Lanman, Daniel Crispell, and Gabriel Taubin Brown University, Dept. of Engineering August 21, 2007 1 Outline Introduction and Related Work

More information

CS Decision Trees / Random Forests

CS Decision Trees / Random Forests CS548 2015 Decision Trees / Random Forests Showcase by: Lily Amadeo, Bir B Kafle, Suman Kumar Lama, Cody Olivier Showcase work by Jamie Shotton, Andrew Fitzgibbon, Richard Moore, Mat Cook, Alex Kipman,

More information

3D Fusion of Infrared Images with Dense RGB Reconstruction from Multiple Views - with Application to Fire-fighting Robots

3D Fusion of Infrared Images with Dense RGB Reconstruction from Multiple Views - with Application to Fire-fighting Robots 3D Fusion of Infrared Images with Dense RGB Reconstruction from Multiple Views - with Application to Fire-fighting Robots Yuncong Chen 1 and Will Warren 2 1 Department of Computer Science and Engineering,

More information

Occlusion Detection of Real Objects using Contour Based Stereo Matching

Occlusion Detection of Real Objects using Contour Based Stereo Matching Occlusion Detection of Real Objects using Contour Based Stereo Matching Kenichi Hayashi, Hirokazu Kato, Shogo Nishida Graduate School of Engineering Science, Osaka University,1-3 Machikaneyama-cho, Toyonaka,

More information

A COMPREHENSIVE SIMULATION SOFTWARE FOR TEACHING CAMERA CALIBRATION

A COMPREHENSIVE SIMULATION SOFTWARE FOR TEACHING CAMERA CALIBRATION XIX IMEKO World Congress Fundamental and Applied Metrology September 6 11, 2009, Lisbon, Portugal A COMPREHENSIVE SIMULATION SOFTWARE FOR TEACHING CAMERA CALIBRATION David Samper 1, Jorge Santolaria 1,

More information

Development of a Fall Detection System with Microsoft Kinect

Development of a Fall Detection System with Microsoft Kinect Development of a Fall Detection System with Microsoft Kinect Christopher Kawatsu, Jiaxing Li, and C.J. Chung Department of Mathematics and Computer Science, Lawrence Technological University, 21000 West

More information

Lecture 8 Active stereo & Volumetric stereo

Lecture 8 Active stereo & Volumetric stereo Lecture 8 Active stereo & Volumetric stereo In this lecture, we ll first discuss another framework for describing stereo systems called active stereo, and then introduce the problem of volumetric stereo,

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

Handheld scanning with ToF sensors and cameras

Handheld scanning with ToF sensors and cameras Handheld scanning with ToF sensors and cameras Enrico Cappelletto, Pietro Zanuttigh, Guido Maria Cortelazzo Dept. of Information Engineering, University of Padova enrico.cappelletto,zanuttigh,corte@dei.unipd.it

More information

Camera Calibration for a Robust Omni-directional Photogrammetry System

Camera Calibration for a Robust Omni-directional Photogrammetry System Camera Calibration for a Robust Omni-directional Photogrammetry System Fuad Khan 1, Michael Chapman 2, Jonathan Li 3 1 Immersive Media Corporation Calgary, Alberta, Canada 2 Ryerson University Toronto,

More information

Computer and Machine Vision

Computer and Machine Vision Computer and Machine Vision Lecture Week 12 Part-2 Additional 3D Scene Considerations March 29, 2014 Sam Siewert Outline of Week 12 Computer Vision APIs and Languages Alternatives to C++ and OpenCV API

More information

FAST HUMAN DETECTION USING TEMPLATE MATCHING FOR GRADIENT IMAGES AND ASC DESCRIPTORS BASED ON SUBTRACTION STEREO

FAST HUMAN DETECTION USING TEMPLATE MATCHING FOR GRADIENT IMAGES AND ASC DESCRIPTORS BASED ON SUBTRACTION STEREO FAST HUMAN DETECTION USING TEMPLATE MATCHING FOR GRADIENT IMAGES AND ASC DESCRIPTORS BASED ON SUBTRACTION STEREO Makoto Arie, Masatoshi Shibata, Kenji Terabayashi, Alessandro Moro and Kazunori Umeda Course

More information

Data-driven Depth Inference from a Single Still Image

Data-driven Depth Inference from a Single Still Image Data-driven Depth Inference from a Single Still Image Kyunghee Kim Computer Science Department Stanford University kyunghee.kim@stanford.edu Abstract Given an indoor image, how to recover its depth information

More information

ENGN D Photography / Spring 2018 / SYLLABUS

ENGN D Photography / Spring 2018 / SYLLABUS ENGN 2502 3D Photography / Spring 2018 / SYLLABUS Description of the proposed course Over the last decade digital photography has entered the mainstream with inexpensive, miniaturized cameras routinely

More information

Improved Navigated Spine Surgery Utilizing Augmented Reality Visualization

Improved Navigated Spine Surgery Utilizing Augmented Reality Visualization Improved Navigated Spine Surgery Utilizing Augmented Reality Visualization Zein Salah 1,2, Bernhard Preim 1, Erck Elolf 3, Jörg Franke 4, Georg Rose 2 1Department of Simulation and Graphics, University

More information

HISTOGRAMS OF ORIENTATIO N GRADIENTS

HISTOGRAMS OF ORIENTATIO N GRADIENTS HISTOGRAMS OF ORIENTATIO N GRADIENTS Histograms of Orientation Gradients Objective: object recognition Basic idea Local shape information often well described by the distribution of intensity gradients

More information

3D Reconstruction of a Hopkins Landmark

3D Reconstruction of a Hopkins Landmark 3D Reconstruction of a Hopkins Landmark Ayushi Sinha (461), Hau Sze (461), Diane Duros (361) Abstract - This paper outlines a method for 3D reconstruction from two images. Our procedure is based on known

More information

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems EECS 442 Computer vision Stereo systems Stereo vision Rectification Correspondence problem Active stereo vision systems Reading: [HZ] Chapter: 11 [FP] Chapter: 11 Stereo vision P p p O 1 O 2 Goal: estimate

More information

Human Motion Detection and Tracking for Video Surveillance

Human Motion Detection and Tracking for Video Surveillance Human Motion Detection and Tracking for Video Surveillance Prithviraj Banerjee and Somnath Sengupta Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur,

More information

Stereoscopic Vision System for reconstruction of 3D objects

Stereoscopic Vision System for reconstruction of 3D objects Stereoscopic Vision System for reconstruction of 3D objects Robinson Jimenez-Moreno Professor, Department of Mechatronics Engineering, Nueva Granada Military University, Bogotá, Colombia. Javier O. Pinzón-Arenas

More information

Vehicle Dimensions Estimation Scheme Using AAM on Stereoscopic Video

Vehicle Dimensions Estimation Scheme Using AAM on Stereoscopic Video Workshop on Vehicle Retrieval in Surveillance (VRS) in conjunction with 2013 10th IEEE International Conference on Advanced Video and Signal Based Surveillance Vehicle Dimensions Estimation Scheme Using

More information

Outline. ETN-FPI Training School on Plenoptic Sensing

Outline. ETN-FPI Training School on Plenoptic Sensing Outline Introduction Part I: Basics of Mathematical Optimization Linear Least Squares Nonlinear Optimization Part II: Basics of Computer Vision Camera Model Multi-Camera Model Multi-Camera Calibration

More information

#65 MONITORING AND PREDICTING PEDESTRIAN BEHAVIOR AT TRAFFIC INTERSECTIONS

#65 MONITORING AND PREDICTING PEDESTRIAN BEHAVIOR AT TRAFFIC INTERSECTIONS #65 MONITORING AND PREDICTING PEDESTRIAN BEHAVIOR AT TRAFFIC INTERSECTIONS Final Research Report Luis E. Navarro-Serment, Ph.D. The Robotics Institute Carnegie Mellon University Disclaimer The contents

More information

Supplementary Material: Decision Tree Fields

Supplementary Material: Decision Tree Fields Supplementary Material: Decision Tree Fields Note, the supplementary material is not needed to understand the main paper. Sebastian Nowozin Sebastian.Nowozin@microsoft.com Toby Sharp toby.sharp@microsoft.com

More information

Dynamic Time Warping for Binocular Hand Tracking and Reconstruction

Dynamic Time Warping for Binocular Hand Tracking and Reconstruction Dynamic Time Warping for Binocular Hand Tracking and Reconstruction Javier Romero, Danica Kragic Ville Kyrki Antonis Argyros CAS-CVAP-CSC Dept. of Information Technology Institute of Computer Science KTH,

More information

3D-2D Laser Range Finder calibration using a conic based geometry shape

3D-2D Laser Range Finder calibration using a conic based geometry shape 3D-2D Laser Range Finder calibration using a conic based geometry shape Miguel Almeida 1, Paulo Dias 1, Miguel Oliveira 2, Vítor Santos 2 1 Dept. of Electronics, Telecom. and Informatics, IEETA, University

More information

3D Object Model Acquisition from Silhouettes

3D Object Model Acquisition from Silhouettes 4th International Symposium on Computing and Multimedia Studies 1 3D Object Model Acquisition from Silhouettes Masaaki Iiyama Koh Kakusho Michihiko Minoh Academic Center for Computing and Media Studies

More information

Tracking Under Low-light Conditions Using Background Subtraction

Tracking Under Low-light Conditions Using Background Subtraction Tracking Under Low-light Conditions Using Background Subtraction Matthew Bennink Clemson University Clemson, South Carolina Abstract A low-light tracking system was developed using background subtraction.

More information

The main problem of photogrammetry

The main problem of photogrammetry Structured Light Structured Light The main problem of photogrammetry to recover shape from multiple views of a scene, we need to find correspondences between the images the matching/correspondence problem

More information

Archeoviz: Improving the Camera Calibration Process. Jonathan Goulet Advisor: Dr. Kostas Daniilidis

Archeoviz: Improving the Camera Calibration Process. Jonathan Goulet Advisor: Dr. Kostas Daniilidis Archeoviz: Improving the Camera Calibration Process Jonathan Goulet Advisor: Dr. Kostas Daniilidis Problem Project Description Complete 3-D reconstruction of site in Tiwanaku, Bolivia Program for archeologists

More information

STEREO VISION AND LASER STRIPERS FOR THREE-DIMENSIONAL SURFACE MEASUREMENTS

STEREO VISION AND LASER STRIPERS FOR THREE-DIMENSIONAL SURFACE MEASUREMENTS XVI CONGRESO INTERNACIONAL DE INGENIERÍA GRÁFICA STEREO VISION AND LASER STRIPERS FOR THREE-DIMENSIONAL SURFACE MEASUREMENTS BARONE, Sandro; BRUNO, Andrea University of Pisa Dipartimento di Ingegneria

More information

Fully Automatic Endoscope Calibration for Intraoperative Use

Fully Automatic Endoscope Calibration for Intraoperative Use Fully Automatic Endoscope Calibration for Intraoperative Use Christian Wengert, Mireille Reeff, Philippe C. Cattin, Gábor Székely Computer Vision Laboratory, ETH Zurich, 8092 Zurich, Switzerland {wengert,

More information

Horus: Object Orientation and Id without Additional Markers

Horus: Object Orientation and Id without Additional Markers Computer Science Department of The University of Auckland CITR at Tamaki Campus (http://www.citr.auckland.ac.nz) CITR-TR-74 November 2000 Horus: Object Orientation and Id without Additional Markers Jacky

More information

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006 Camera Registration in a 3D City Model Min Ding CS294-6 Final Presentation Dec 13, 2006 Goal: Reconstruct 3D city model usable for virtual walk- and fly-throughs Virtual reality Urban planning Simulation

More information

REPRESENTATION REQUIREMENTS OF AS-IS BUILDING INFORMATION MODELS GENERATED FROM LASER SCANNED POINT CLOUD DATA

REPRESENTATION REQUIREMENTS OF AS-IS BUILDING INFORMATION MODELS GENERATED FROM LASER SCANNED POINT CLOUD DATA REPRESENTATION REQUIREMENTS OF AS-IS BUILDING INFORMATION MODELS GENERATED FROM LASER SCANNED POINT CLOUD DATA Engin Burak Anil 1 *, Burcu Akinci 1, and Daniel Huber 2 1 Department of Civil and Environmental

More information

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,

More information

Stereo Image Rectification for Simple Panoramic Image Generation

Stereo Image Rectification for Simple Panoramic Image Generation Stereo Image Rectification for Simple Panoramic Image Generation Yun-Suk Kang and Yo-Sung Ho Gwangju Institute of Science and Technology (GIST) 261 Cheomdan-gwagiro, Buk-gu, Gwangju 500-712 Korea Email:{yunsuk,

More information

Measurement of Pedestrian Groups Using Subtraction Stereo

Measurement of Pedestrian Groups Using Subtraction Stereo Measurement of Pedestrian Groups Using Subtraction Stereo Kenji Terabayashi, Yuki Hashimoto, and Kazunori Umeda Chuo University / CREST, JST, 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551, Japan terabayashi@mech.chuo-u.ac.jp

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Omni-directional Multi-baseline Stereo without Similarity Measures

Omni-directional Multi-baseline Stereo without Similarity Measures Omni-directional Multi-baseline Stereo without Similarity Measures Tomokazu Sato and Naokazu Yokoya Graduate School of Information Science, Nara Institute of Science and Technology 8916-5 Takayama, Ikoma,

More information

International Conference on Communication, Media, Technology and Design. ICCMTD May 2012 Istanbul - Turkey

International Conference on Communication, Media, Technology and Design. ICCMTD May 2012 Istanbul - Turkey VISUALIZING TIME COHERENT THREE-DIMENSIONAL CONTENT USING ONE OR MORE MICROSOFT KINECT CAMERAS Naveed Ahmed University of Sharjah Sharjah, United Arab Emirates Abstract Visualizing or digitization of the

More information

Project Updates Short lecture Volumetric Modeling +2 papers

Project Updates Short lecture Volumetric Modeling +2 papers Volumetric Modeling Schedule (tentative) Feb 20 Feb 27 Mar 5 Introduction Lecture: Geometry, Camera Model, Calibration Lecture: Features, Tracking/Matching Mar 12 Mar 19 Mar 26 Apr 2 Apr 9 Apr 16 Apr 23

More information

Mapping Non-Destructive Testing Data on the 3D Geometry of Objects with Complex Shapes

Mapping Non-Destructive Testing Data on the 3D Geometry of Objects with Complex Shapes More Info at Open Access Database www.ndt.net/?id=17670 Mapping Non-Destructive Testing Data on the 3D Geometry of Objects with Complex Shapes Abstract by S. Soldan*, D. Ouellet**, P.Hedayati**, A. Bendada**,

More information

Transparent Object Shape Measurement Based on Deflectometry

Transparent Object Shape Measurement Based on Deflectometry Proceedings Transparent Object Shape Measurement Based on Deflectometry Zhichao Hao and Yuankun Liu * Opto-Electronics Department, Sichuan University, Chengdu 610065, China; 2016222055148@stu.scu.edu.cn

More information

ENGN2911I: 3D Photography and Geometry Processing Assignment 1: 3D Photography using Planar Shadows

ENGN2911I: 3D Photography and Geometry Processing Assignment 1: 3D Photography using Planar Shadows ENGN2911I: 3D Photography and Geometry Processing Assignment 1: 3D Photography using Planar Shadows Instructor: Gabriel Taubin Assignment written by: Douglas Lanman 29 January 2009 Figure 1: 3D Photography

More information

EECS 442 Computer vision. Announcements

EECS 442 Computer vision. Announcements EECS 442 Computer vision Announcements Midterm released after class (at 5pm) You ll have 46 hours to solve it. it s take home; you can use your notes and the books no internet must work on it individually

More information

Easy to Use Calibration of Multiple Camera Setups

Easy to Use Calibration of Multiple Camera Setups Easy to Use Calibration of Multiple Camera Setups Ferenc Kahlesz, Cornelius Lilge, and Reinhard Klein University of Bonn, Institute of Computer Science II, Computer Graphics Group Römerstrasse 164, D-53117

More information

Calibration of a Different Field-of-view Stereo Camera System using an Embedded Checkerboard Pattern

Calibration of a Different Field-of-view Stereo Camera System using an Embedded Checkerboard Pattern Calibration of a Different Field-of-view Stereo Camera System using an Embedded Checkerboard Pattern Pathum Rathnayaka, Seung-Hae Baek and Soon-Yong Park School of Computer Science and Engineering, Kyungpook

More information

Real Time Motion Detection Using Background Subtraction Method and Frame Difference

Real Time Motion Detection Using Background Subtraction Method and Frame Difference Real Time Motion Detection Using Background Subtraction Method and Frame Difference Lavanya M P PG Scholar, Department of ECE, Channabasaveshwara Institute of Technology, Gubbi, Tumkur Abstract: In today

More information

Multiple View Geometry

Multiple View Geometry Multiple View Geometry CS 6320, Spring 2013 Guest Lecture Marcel Prastawa adapted from Pollefeys, Shah, and Zisserman Single view computer vision Projective actions of cameras Camera callibration Photometric

More information

Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera

Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera Wolfgang Niem, Jochen Wingbermühle Universität Hannover Institut für Theoretische Nachrichtentechnik und Informationsverarbeitung

More information

DTU Technical Report: ARTTS

DTU Technical Report: ARTTS DTU Technical Report: ARTTS Title: Author: Project: Face pose tracking and recognition and 3D cameras Rasmus Larsen ARTTS Date: February 10 th, 2006 Contents Contents...2 Introduction...2 State-of-the

More information

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION

Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen I. INTRODUCTION Kinect Cursor Control EEE178 Dr. Fethi Belkhouche Christopher Harris Danny Nguyen Abstract: An XBOX 360 Kinect is used to develop two applications to control the desktop cursor of a Windows computer. Application

More information

3D Digitization of a Hand-held Object with a Wearable Vision Sensor

3D Digitization of a Hand-held Object with a Wearable Vision Sensor 3D Digitization of a Hand-held Object with a Wearable Vision Sensor Sotaro TSUKIZAWA, Kazuhiko SUMI, and Takashi MATSUYAMA tsucky@vision.kuee.kyoto-u.ac.jp sumi@vision.kuee.kyoto-u.ac.jp tm@i.kyoto-u.ac.jp

More information

Light source estimation using feature points from specular highlights and cast shadows

Light source estimation using feature points from specular highlights and cast shadows Vol. 11(13), pp. 168-177, 16 July, 2016 DOI: 10.5897/IJPS2015.4274 Article Number: F492B6D59616 ISSN 1992-1950 Copyright 2016 Author(s) retain the copyright of this article http://www.academicjournals.org/ijps

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

LUMS Mine Detector Project

LUMS Mine Detector Project LUMS Mine Detector Project Using visual information to control a robot (Hutchinson et al. 1996). Vision may or may not be used in the feedback loop. Visual (image based) features such as points, lines

More information

of human activities. Our research is motivated by considerations of a ground-based mobile surveillance system that monitors an extended area for

of human activities. Our research is motivated by considerations of a ground-based mobile surveillance system that monitors an extended area for To Appear in ACCV-98, Mumbai-India, Material Subject to ACCV Copy-Rights Visual Surveillance of Human Activity Larry Davis 1 Sandor Fejes 1 David Harwood 1 Yaser Yacoob 1 Ismail Hariatoglu 1 Michael J.

More information

Dynamic Human Shape Description and Characterization

Dynamic Human Shape Description and Characterization Dynamic Human Shape Description and Characterization Z. Cheng*, S. Mosher, Jeanne Smith H. Cheng, and K. Robinette Infoscitex Corporation, Dayton, Ohio, USA 711 th Human Performance Wing, Air Force Research

More information

People detection and tracking using stereo vision and color

People detection and tracking using stereo vision and color People detection and tracking using stereo vision and color Rafael Munoz-Salinas, Eugenio Aguirre, Miguel Garcia-Silvente. In Image and Vision Computing Volume 25 Issue 6 (2007) 995-1007. Presented by

More information

Multi-View Stereo for Static and Dynamic Scenes

Multi-View Stereo for Static and Dynamic Scenes Multi-View Stereo for Static and Dynamic Scenes Wolfgang Burgard Jan 6, 2010 Main references Yasutaka Furukawa and Jean Ponce, Accurate, Dense and Robust Multi-View Stereopsis, 2007 C.L. Zitnick, S.B.

More information

Gregory Walsh, Ph.D. San Ramon, CA January 25, 2011

Gregory Walsh, Ph.D. San Ramon, CA January 25, 2011 Leica ScanStation:: Calibration and QA Gregory Walsh, Ph.D. San Ramon, CA January 25, 2011 1. Summary Leica Geosystems, in creating the Leica Scanstation family of products, has designed and conducted

More information

HIGH SPEED 3-D MEASUREMENT SYSTEM USING INCOHERENT LIGHT SOURCE FOR HUMAN PERFORMANCE ANALYSIS

HIGH SPEED 3-D MEASUREMENT SYSTEM USING INCOHERENT LIGHT SOURCE FOR HUMAN PERFORMANCE ANALYSIS HIGH SPEED 3-D MEASUREMENT SYSTEM USING INCOHERENT LIGHT SOURCE FOR HUMAN PERFORMANCE ANALYSIS Takeo MIYASAKA, Kazuhiro KURODA, Makoto HIROSE and Kazuo ARAKI School of Computer and Cognitive Sciences,

More information

And. Modal Analysis. Using. VIC-3D-HS, High Speed 3D Digital Image Correlation System. Indian Institute of Technology New Delhi

And. Modal Analysis. Using. VIC-3D-HS, High Speed 3D Digital Image Correlation System. Indian Institute of Technology New Delhi Full Field Displacement And Strain Measurement And Modal Analysis Using VIC-3D-HS, High Speed 3D Digital Image Correlation System At Indian Institute of Technology New Delhi VIC-3D, 3D Digital Image Correlation

More information

Depth Range Accuracy for Plenoptic Cameras

Depth Range Accuracy for Plenoptic Cameras Depth Range Accuracy for Plenoptic Cameras Nuno Barroso Monteiro Institute for Systems and Robotics, University of Lisbon, Portugal Institute for Systems and Robotics, University of Coimbra, Portugal Simão

More information

Gait analysis for person recognition using principal component analysis and support vector machines

Gait analysis for person recognition using principal component analysis and support vector machines Gait analysis for person recognition using principal component analysis and support vector machines O V Strukova 1, LV Shiripova 1 and E V Myasnikov 1 1 Samara National Research University, Moskovskoe

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

Structured light 3D reconstruction

Structured light 3D reconstruction Structured light 3D reconstruction Reconstruction pipeline and industrial applications rodola@dsi.unive.it 11/05/2010 3D Reconstruction 3D reconstruction is the process of capturing the shape and appearance

More information

Markerless human motion capture through visual hull and articulated ICP

Markerless human motion capture through visual hull and articulated ICP Markerless human motion capture through visual hull and articulated ICP Lars Mündermann lmuender@stanford.edu Stefano Corazza Stanford, CA 93405 stefanoc@stanford.edu Thomas. P. Andriacchi Bone and Joint

More information

Temporally-Consistent Phase Unwrapping for a Stereo-Assisted Structured Light System

Temporally-Consistent Phase Unwrapping for a Stereo-Assisted Structured Light System Temporally-Consistent Phase Unwrapping for a Stereo-Assisted Structured Light System Ricardo R. Garcia and Avideh Zakhor Department of Electrical Engineering and Computer Science University of California,

More information

Image Based Reconstruction II

Image Based Reconstruction II Image Based Reconstruction II Qixing Huang Feb. 2 th 2017 Slide Credit: Yasutaka Furukawa Image-Based Geometry Reconstruction Pipeline Last Lecture: Multi-View SFM Multi-View SFM This Lecture: Multi-View

More information

3D Modeling of Objects Using Laser Scanning

3D Modeling of Objects Using Laser Scanning 1 3D Modeling of Objects Using Laser Scanning D. Jaya Deepu, LPU University, Punjab, India Email: Jaideepudadi@gmail.com Abstract: In the last few decades, constructing accurate three-dimensional models

More information

Miniaturized Camera Systems for Microfactories

Miniaturized Camera Systems for Microfactories Miniaturized Camera Systems for Microfactories Timo Prusi, Petri Rokka, and Reijo Tuokko Tampere University of Technology, Department of Production Engineering, Korkeakoulunkatu 6, 33720 Tampere, Finland

More information

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe Structured Light Tobias Nöll tobias.noell@dfki.de Thanks to Marc Pollefeys, David Nister and David Lowe Introduction Previous lecture: Dense reconstruction Dense matching of non-feature pixels Patch-based

More information

Reduced Image Noise on Shape Recognition Using Singular Value Decomposition for Pick and Place Robotic Systems

Reduced Image Noise on Shape Recognition Using Singular Value Decomposition for Pick and Place Robotic Systems Reduced Image Noise on Shape Recognition Using Singular Value Decomposition for Pick and Place Robotic Systems Angelo A. Beltran Jr. 1, Christian Deus T. Cayao 2, Jay-K V. Delicana 3, Benjamin B. Agraan

More information

3D Vision Real Objects, Real Cameras. Chapter 11 (parts of), 12 (parts of) Computerized Image Analysis MN2 Anders Brun,

3D Vision Real Objects, Real Cameras. Chapter 11 (parts of), 12 (parts of) Computerized Image Analysis MN2 Anders Brun, 3D Vision Real Objects, Real Cameras Chapter 11 (parts of), 12 (parts of) Computerized Image Analysis MN2 Anders Brun, anders@cb.uu.se 3D Vision! Philisophy! Image formation " The pinhole camera " Projective

More information

Video Processing for Judicial Applications

Video Processing for Judicial Applications Video Processing for Judicial Applications Konstantinos Avgerinakis, Alexia Briassouli, Ioannis Kompatsiaris Informatics and Telematics Institute, Centre for Research and Technology, Hellas Thessaloniki,

More information

Part I: HumanEva-I dataset and evaluation metrics

Part I: HumanEva-I dataset and evaluation metrics Part I: HumanEva-I dataset and evaluation metrics Leonid Sigal Michael J. Black Department of Computer Science Brown University http://www.cs.brown.edu/people/ls/ http://vision.cs.brown.edu/humaneva/ Motivation

More information

Face Detection Using Convolutional Neural Networks and Gabor Filters

Face Detection Using Convolutional Neural Networks and Gabor Filters Face Detection Using Convolutional Neural Networks and Gabor Filters Bogdan Kwolek Rzeszów University of Technology W. Pola 2, 35-959 Rzeszów, Poland bkwolek@prz.rzeszow.pl Abstract. This paper proposes

More information

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg

Human Detection. A state-of-the-art survey. Mohammad Dorgham. University of Hamburg Human Detection A state-of-the-art survey Mohammad Dorgham University of Hamburg Presentation outline Motivation Applications Overview of approaches (categorized) Approaches details References Motivation

More information

Mixed-Reality for Intuitive Photo-Realistic 3D-Model Generation

Mixed-Reality for Intuitive Photo-Realistic 3D-Model Generation Mixed-Reality for Intuitive Photo-Realistic 3D-Model Generation Wolfgang Sepp, Tim Bodenmueller, Michael Suppa, and Gerd Hirzinger DLR, Institut für Robotik und Mechatronik @ GI-Workshop VR/AR 2009 Folie

More information

Self-learning Voxel-based Multi-camera Occlusion Maps for 3D Reconstruction

Self-learning Voxel-based Multi-camera Occlusion Maps for 3D Reconstruction Self-learning Voxel-based Multi-camera Occlusion Maps for 3D Reconstruction Maarten Slembrouck 1, Dimitri Van Cauwelaert 1, David Van Hamme 1, Dirk Van Haerenborgh 1, Peter Van Hese 1, Peter Veelaert 1

More information

Multiple View Geometry

Multiple View Geometry Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V

More information