ATIP A Tool for 3D Navigation inside a Single Image with Automatic Camera Calibration
|
|
- Nelson Carson
- 5 years ago
- Views:
Transcription
1 ATIP A Tool for 3D Navigation inside a Single Image with Automatic Camera Calibration Kévin Boulanger, Kadi Bouatouch, Sumanta Pattanaik IRISA, Université de Rennes I, France University of Central Florida, USA Eurographics UK Theory and Practice of Computer Graphics 2006
2 Purpose Single image Simple 3D navigation 2
3 Software architecture ATIP Maker : processes the input image, creates an.atip file ATIP Navigator : reads the.atip file, allows multi-platform 3D navigation Output Input image.atip file ATIP ATIP maker maker ATIP ATIP navigator navigator 3
4 Tour Into the Picture Original method by Horry et al. Given an input image, the perspective effect is determined by manually (mouse click) choosing a point in the image as a vanishing point The scene is manually approximated by a box by dragging its corner handles Textures are computed for each box s face from the input image The textured box is rendered from another viewpoint box with handles vanishing point 4
5 Tour Into the Picture Manual fitting of the box long and tedious Limitations supports only images with a single vanishing point horizontal and vertical lines must be parallel to the image borders careful capture of the input image is necessary (say: using a tripod) 5
6 Automatic Tour Into the Picture Contributions of ATIP: minimal user interaction more accurate calibration of the camera using multiple vanishing points input image from any orientation of the camera is acceptable the camera can be hand-held 6
7 Vanishing points Vanishing points help to calibrate the camera Always three vanishing points in an image Three combinations of finite and infinite vanishing points TIP manages only one finite vanishing point ATIP manages every situation, even with rotation around the view direction (very common without tripod) Horizontal IVP FVP Vertical IVP Vertical IVP FVP FVP FVP FVP FVP 1 finite vanishing point, 2 infinite vanishing points Horizon line 2 finite vanishing points, 1 infinite vanishing point 3 finite vanishing points 7
8 ATIP Maker pipeline First part: image processing Input image Edge detection Dominant lines detection Vanishing points estimation Three vanishing points 8
9 ATIP Maker pipeline Second part: determining parameters for TIP (camera, box, textures) Three vanishing points Camera parameter extraction Box size adjustment Texture computation Textures Geometry Input image 9
10 Edge detection Conversion to grayscale of the input image Logical AND between gradient-based and laplacian-based methods Input image Grayscale image Edges 10
11 Dominant lines detection Dominant lines: longest alignments of points Vanishing lines = subset of the set of dominant lines How: 1-to-m Hough transform of the edge points high-pass filter of the Hough transform result detection of the maxima by thresholding transform the maxima back to lines in image space Edges Hough transform High-pass filter Dominant lines 11
12 Dominant lines detection 1-to-m Hough transform Image space y Hough space Line (with polar parametrization) ρ θ x ρ ρ θ Point Point (intersection of an infinite number of lines) y θ Curve ρ = x.cos(θ) + y.sin(θ) y x ρ Aligned points Set of curves with a common point, θ corresponding to a set of aligned points x in image space 12
13 Dominant lines detection Hough space high-pass filter Simple thresholding of the Hough transform lot of noise due to strongly textured regions of the input image High-pass filter to remove the dense regions of high values (noise) Region with points to be kept Kept points Dense region of high values to be eliminated by filtering Noise eliminated Hough transform Simple thresholding High-pass filter Final result 13
14 Vanishing points estimation Vanishing point = intersection of a subset of dominant lines Projection of each dominant line onto a hemisphere Image plane Line in image space to project Accumulation of the resulting curves on the hemisphere Projection of the 2D line θ φ x Retrieve the three maxima y Project them back to image space to get the vanishing point coordinates z 14
15 Vanishing points estimation Accumulation The intersection of curves on the hemisphere corresponds to the intersection of the corresponding lines in image space (pole) Grid of accumulators i )) Finite vanishing point (close to the image center) Lines to project Image plane Final intersection point (equator) N/4 N/2 3N/4 N-1 i y z Projection of one line Intersection point of the circles x Horizontal infinite vanishing point C θ φ (x,y,z) Vertical infinite vanishing point Strips of equal area x Non-uniform subdivision of the hemisphere y z 15
16 Vanishing points estimation Three maxima retrieval The three maxima cannot be retrieved in a single step, instead : for i = 0 to 2 { filteredcells = low_pass_filter(hemisphere_cells); maxcell = maximum_value_cell(filteredcells); vanpoints[i] = project_to_image_plane(maxcell); dominantlines[i] = dominant_lines_associated_with(vanpoints[i]); hemisphere_cells = negative_accumulation(dominantlines[i], hemisphere_cells) } Hypothesis: the three resulting vanishing points correspond to orthogonal directions With this algorithm, only dominant lines that contribute to the detection of vanishing points are considered 16
17 Camera parameter extraction From the three vanishing points, finite or infinite, extract camera parameters Focal length and rotation matrix The three vanishing point directions define the world coordinate frame Process the three combinations of vanishing points different ways 17
18 Camera parameter extraction One finite VP, two infinite VPs rv 1 finite r vanishing point I 1 and I 2 directions of the infinite vanishing points in image space O position of the camera x r, y r, z r coordinate frame of the camera (fixed) (focal length) is fixed by the user (48 by default) u r f, v r, w r world coordinate frame (columns of the rotation matrix), to be found r I I I, 0 I r OV = V V, f I I, 0 1 ( ) T 1x, 1y 1 = ( ) T 1 x, 1 y 2 = ( ) T, 2x 2 y image plane V 1 I 1 P f I 2 O y v u x w z r I1xV1 x + I u = I1 x, I1y, f 1y r I 2xV1 x + I v = I 2 x, I 2 y, f ( V V f ) T r w =, 1x, 1y V 2 y 1y V 1y T T r u = r v = r w = u r r u v r r v w r r w 18
19 Camera parameter extraction Two finite VPs, one infinite VP Two known VPs ( 1 and ), represent orthogonal directions f can be computed obtained by cross product v r V V2 OV OV 1 2 ( V ) T 1x, V1 y f ( V V f ) T =, =, 2x, 2 y image plane V 2 V 1 OV1. OV2 = 0 f > 0 P f = V1 xv1 y + V2 xv2 y f x u y O v w z r u = OV OV r r r v = w u 1 1 r w = OV OV
20 Camera parameter extraction Three finite vanishing points v r 1 2 Three finite vanishing points,, and overconstrained problem The two finite VPs method used for ( v r, v r ) 1 2, ( v r, r ) and ( v r, r ) 2 v 3 3 v 1 The dot product between the third VP direction and the cross product of the two first directions gives an error value Choose the set of vectors giving minimal error v r v r 3 20
21 Box size adjustment The only step of the algorithm that requires user interaction Simpler than TIP since the camera is calibrated, only the corners (red handles) have to be moved 21
22 Texture computation Data retrieved from the input image Each box s face is subdivided into a uniform grid, each cell corresponds to a texel Each texel center point of a texture is projected onto the input image plane Bilinear interpolation gives the final color Points falling outside the input image bounds are set to black textured quadrilateral (x,y,z) image plane (x i,y i ) O 22
23 Results 23
24 Results 24
25 Results with difficult scenes Forest scene: Robust dominant lines detection from noisy edges using high-pass filter on Hough transform Non-flat surfaces: Robust «maxima on the hemisphere» computation algorithm Hand-drawn sketch: Correct detection of dominant lines from noisy edges and imprecise intersections of vanishing lines 25
26 Thanks for your attention Questions? For more information, visit
ATIP: A Tool for 3D Navigation inside a Single Image with Automatic Camera Calibration
EG UK Theory and Practice of Computer Graphics (2006) M. McDerby, L. Lever (Editors) ATIP: A Tool for 3D Navigation inside a Single Image with Automatic Camera Calibration K. Boulanger, K. Bouatouch and
More informationRendering Grass Terrains in Real-Time with Dynamic Lighting. Kévin Boulanger, Sumanta Pattanaik, Kadi Bouatouch August 1st 2006
Rendering Grass Terrains in Real-Time with Dynamic Lighting Kévin Boulanger, Sumanta Pattanaik, Kadi Bouatouch August 1st 2006 Goal Rendering millions of grass blades, at any distance, in real-time, with:
More informationScene Modeling for a Single View
Scene Modeling for a Single View René MAGRITTE Portrait d'edward James with a lot of slides stolen from Steve Seitz and David Brogan, Breaking out of 2D now we are ready to break out of 2D And enter the
More informationEdge linking. Two types of approaches. This process needs to be able to bridge gaps in detected edges due to the reason mentioned above
Edge linking Edge detection rarely finds the entire set of edges in an image. Normally there are breaks due to noise, non-uniform illumination, etc. If we want to obtain region boundaries (for segmentation)
More informationScene Modeling for a Single View
Scene Modeling for a Single View René MAGRITTE Portrait d'edward James with a lot of slides stolen from Steve Seitz and David Brogan, 15-463: Computational Photography Alexei Efros, CMU, Fall 2005 Classes
More informationScene Modeling for a Single View
Scene Modeling for a Single View René MAGRITTE Portrait d'edward James CS194: Image Manipulation & Computational Photography with a lot of slides stolen from Alexei Efros, UC Berkeley, Fall 2014 Steve
More information2D rendering takes a photo of the 2D scene with a virtual camera that selects an axis aligned rectangle from the scene. The photograph is placed into
2D rendering takes a photo of the 2D scene with a virtual camera that selects an axis aligned rectangle from the scene. The photograph is placed into the viewport of the current application window. A pixel
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 10 130221 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Canny Edge Detector Hough Transform Feature-Based
More informationLecture 9: Hough Transform and Thresholding base Segmentation
#1 Lecture 9: Hough Transform and Thresholding base Segmentation Saad Bedros sbedros@umn.edu Hough Transform Robust method to find a shape in an image Shape can be described in parametric form A voting
More informationHOUGH TRANSFORM. Plan for today. Introduction to HT. An image with linear structures. INF 4300 Digital Image Analysis
INF 4300 Digital Image Analysis HOUGH TRANSFORM Fritz Albregtsen 14.09.2011 Plan for today This lecture goes more in detail than G&W 10.2! Introduction to Hough transform Using gradient information to
More informationSingle-view 3D Reconstruction
Single-view 3D Reconstruction 10/12/17 Computational Photography Derek Hoiem, University of Illinois Some slides from Alyosha Efros, Steve Seitz Notes about Project 4 (Image-based Lighting) You can work
More informationScene Modeling for a Single View
on to 3D Scene Modeling for a Single View We want real 3D scene walk-throughs: rotation translation Can we do it from a single photograph? Reading: A. Criminisi, I. Reid and A. Zisserman, Single View Metrology
More informationElaborazione delle Immagini Informazione Multimediale. Raffaella Lanzarotti
Elaborazione delle Immagini Informazione Multimediale Raffaella Lanzarotti HOUGH TRANSFORM Paragraph 4.3.2 of the book at link: szeliski.org/book/drafts/szeliskibook_20100903_draft.pdf Thanks to Kristen
More informationCS 130 Final. Fall 2015
CS 130 Final Fall 2015 Name Student ID Signature You may not ask any questions during the test. If you believe that there is something wrong with a question, write down what you think the question is trying
More information(Refer Slide Time: 0:32)
Digital Image Processing. Professor P. K. Biswas. Department of Electronics and Electrical Communication Engineering. Indian Institute of Technology, Kharagpur. Lecture-57. Image Segmentation: Global Processing
More informationCSCI 5980: Assignment #3 Homography
Submission Assignment due: Feb 23 Individual assignment. Write-up submission format: a single PDF up to 3 pages (more than 3 page assignment will be automatically returned.). Code and data. Submission
More informationE0005E - Industrial Image Analysis
E0005E - Industrial Image Analysis The Hough Transform Matthew Thurley slides by Johan Carlson 1 This Lecture The Hough transform Detection of lines Detection of other shapes (the generalized Hough transform)
More informationFeature Selection. Ardy Goshtasby Wright State University and Image Fusion Systems Research
Feature Selection Ardy Goshtasby Wright State University and Image Fusion Systems Research Image features Points Lines Regions Templates 2 Corners They are 1) locally unique and 2) rotationally invariant
More informationFitting: The Hough transform
Fitting: The Hough transform Voting schemes Let each feature vote for all the models that are compatible with it Hopefully the noise features will not vote consistently for any single model Missing data
More informationDD2429 Computational Photography :00-19:00
. Examination: DD2429 Computational Photography 202-0-8 4:00-9:00 Each problem gives max 5 points. In order to pass you need about 0-5 points. You are allowed to use the lecture notes and standard list
More informationIntroduction to Computer Vision
Introduction to Computer Vision Michael J. Black Nov 2009 Perspective projection and affine motion Goals Today Perspective projection 3D motion Wed Projects Friday Regularization and robust statistics
More informationOther Linear Filters CS 211A
Other Linear Filters CS 211A Slides from Cornelia Fermüller and Marc Pollefeys Edge detection Convert a 2D image into a set of curves Extracts salient features of the scene More compact than pixels Origin
More informationIssues with Curve Detection Grouping (e.g., the Canny hysteresis thresholding procedure) Model tting They can be performed sequentially or simultaneou
an edge image, nd line or curve segments present Given the image. in Line and Curves Detection 1 Issues with Curve Detection Grouping (e.g., the Canny hysteresis thresholding procedure) Model tting They
More informationCSE528 Computer Graphics: Theory, Algorithms, and Applications
CSE528 Computer Graphics: Theory, Algorithms, and Applications Hong Qin State University of New York at Stony Brook (Stony Brook University) Stony Brook, New York 11794--4400 Tel: (631)632-8450; Fax: (631)632-8334
More informationEdge detection. Gradient-based edge operators
Edge detection Gradient-based edge operators Prewitt Sobel Roberts Laplacian zero-crossings Canny edge detector Hough transform for detection of straight lines Circle Hough Transform Digital Image Processing:
More informationThree-Dimensional Viewing Hearn & Baker Chapter 7
Three-Dimensional Viewing Hearn & Baker Chapter 7 Overview 3D viewing involves some tasks that are not present in 2D viewing: Projection, Visibility checks, Lighting effects, etc. Overview First, set up
More informationCamera Geometry II. COS 429 Princeton University
Camera Geometry II COS 429 Princeton University Outline Projective geometry Vanishing points Application: camera calibration Application: single-view metrology Epipolar geometry Application: stereo correspondence
More informationSports Field Localization
Sports Field Localization Namdar Homayounfar February 5, 2017 1 / 1 Motivation Sports Field Localization: Have to figure out where the field and players are in 3d space in order to make measurements and
More informationFeature Detectors and Descriptors: Corners, Lines, etc.
Feature Detectors and Descriptors: Corners, Lines, etc. Edges vs. Corners Edges = maxima in intensity gradient Edges vs. Corners Corners = lots of variation in direction of gradient in a small neighborhood
More informationSimple 3D Reconstruction of Single Indoor Image with Perspe
Simple 3D Reconstruction of Single Indoor Image with Perspective Cues Jingyuan Huang Bill Cowan University of Waterloo May 26, 2009 1 Introduction 2 Our Method 3 Results and Conclusions Problem Statement
More informationComputer Vision I - Appearance-based Matching and Projective Geometry
Computer Vision I - Appearance-based Matching and Projective Geometry Carsten Rother 05/11/2015 Computer Vision I: Image Formation Process Roadmap for next four lectures Computer Vision I: Image Formation
More informationComputer Vision Project-1
University of Utah, School Of Computing Computer Vision Project- Singla, Sumedha sumedha.singla@utah.edu (00877456 February, 205 Theoretical Problems. Pinhole Camera (a A straight line in the world space
More informationAnnouncements. Edges. Last Lecture. Gradients: Numerical Derivatives f(x) Edge Detection, Lines. Intro Computer Vision. CSE 152 Lecture 10
Announcements Assignment 2 due Tuesday, May 4. Edge Detection, Lines Midterm: Thursday, May 6. Introduction to Computer Vision CSE 152 Lecture 10 Edges Last Lecture 1. Object boundaries 2. Surface normal
More informationDETECTION AND ROBUST ESTIMATION OF CYLINDER FEATURES IN POINT CLOUDS INTRODUCTION
DETECTION AND ROBUST ESTIMATION OF CYLINDER FEATURES IN POINT CLOUDS Yun-Ting Su James Bethel Geomatics Engineering School of Civil Engineering Purdue University 550 Stadium Mall Drive, West Lafayette,
More informationHomework #2 and #3 Due Friday, October 12 th and Friday, October 19 th
Homework #2 and #3 Due Friday, October 12 th and Friday, October 19 th 1. a. Show that the following sequences commute: i. A rotation and a uniform scaling ii. Two rotations about the same axis iii. Two
More informationProjective Texture Mapping with Full Panorama
EUROGRAPHICS 2002 / G. Drettakis and H.-P. Seidel Volume 21 (2002), Number 3 (Guest Editors) Projective Texture Mapping with Full Panorama Dongho Kim and James K. Hahn Department of Computer Science, The
More informationCS443: Digital Imaging and Multimedia Perceptual Grouping Detecting Lines and Simple Curves
CS443: Digital Imaging and Multimedia Perceptual Grouping Detecting Lines and Simple Curves Spring 2008 Ahmed Elgammal Dept. of Computer Science Rutgers University Outlines Perceptual Grouping and Segmentation
More informationMosaics. Today s Readings
Mosaics VR Seattle: http://www.vrseattle.com/ Full screen panoramas (cubic): http://www.panoramas.dk/ Mars: http://www.panoramas.dk/fullscreen3/f2_mars97.html Today s Readings Szeliski and Shum paper (sections
More informationStructured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe
Structured Light Tobias Nöll tobias.noell@dfki.de Thanks to Marc Pollefeys, David Nister and David Lowe Introduction Previous lecture: Dense reconstruction Dense matching of non-feature pixels Patch-based
More informationMERGING POINT CLOUDS FROM MULTIPLE KINECTS. Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia
MERGING POINT CLOUDS FROM MULTIPLE KINECTS Nishant Rai 13th July, 2016 CARIS Lab University of British Columbia Introduction What do we want to do? : Use information (point clouds) from multiple (2+) Kinects
More informationDD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication
DD2423 Image Analysis and Computer Vision IMAGE FORMATION Mårten Björkman Computational Vision and Active Perception School of Computer Science and Communication November 8, 2013 1 Image formation Goal:
More information03 Vector Graphics. Multimedia Systems. 2D and 3D Graphics, Transformations
Multimedia Systems 03 Vector Graphics 2D and 3D Graphics, Transformations Imran Ihsan Assistant Professor, Department of Computer Science Air University, Islamabad, Pakistan www.imranihsan.com Lectures
More informationHOUGH TRANSFORM CS 6350 C V
HOUGH TRANSFORM CS 6350 C V HOUGH TRANSFORM The problem: Given a set of points in 2-D, find if a sub-set of these points, fall on a LINE. Hough Transform One powerful global method for detecting edges
More informationFitting: The Hough transform
Fitting: The Hough transform Voting schemes Let each feature vote for all the models that are compatible with it Hopefully the noise features will not vote consistently for any single model Missing data
More informationFinal Review CMSC 733 Fall 2014
Final Review CMSC 733 Fall 2014 We have covered a lot of material in this course. One way to organize this material is around a set of key equations and algorithms. You should be familiar with all of these,
More informationImages from 3D Creative Magazine. 3D Modelling Systems
Images from 3D Creative Magazine 3D Modelling Systems Contents Reference & Accuracy 3D Primitives Transforms Move (Translate) Rotate Scale Mirror Align 3D Booleans Deforms Bend Taper Skew Twist Squash
More informationMore Texture Mapping. Texture Mapping 1/46
More Texture Mapping Texture Mapping 1/46 Perturbing Normals Texture Mapping 2/46 Perturbing Normals Instead of fetching a texture for color, fetch a new perturbed normal vector Creates the appearance
More informationTopics and things to know about them:
Practice Final CMSC 427 Distributed Tuesday, December 11, 2007 Review Session, Monday, December 17, 5:00pm, 4424 AV Williams Final: 10:30 AM Wednesday, December 19, 2007 General Guidelines: The final will
More informationComputer Vision. Image Segmentation. 10. Segmentation. Computer Engineering, Sejong University. Dongil Han
Computer Vision 10. Segmentation Computer Engineering, Sejong University Dongil Han Image Segmentation Image segmentation Subdivides an image into its constituent regions or objects - After an image has
More informationSingle-view metrology
Single-view metrology Magritte, Personal Values, 952 Many slides from S. Seitz, D. Hoiem Camera calibration revisited What if world coordinates of reference 3D points are not known? We can use scene features
More informationChapter 8 Three-Dimensional Viewing Operations
Projections Chapter 8 Three-Dimensional Viewing Operations Figure 8.1 Classification of planar geometric projections Figure 8.2 Planar projection Figure 8.3 Parallel-oblique projection Figure 8.4 Orthographic
More informationA Continuous 3-D Medial Shape Model with Branching
A Continuous 3-D Medial Shape Model with Branching Timothy B. Terriberry Guido Gerig Outline Introduction The Generic 3-D Medial Axis Review of Subdivision Surfaces Boundary Reconstruction Edge Curves
More informationform are graphed in Cartesian coordinates, and are graphed in Cartesian coordinates.
Plot 3D Introduction Plot 3D graphs objects in three dimensions. It has five basic modes: 1. Cartesian mode, where surfaces defined by equations of the form are graphed in Cartesian coordinates, 2. cylindrical
More informationChapter 3 Image Registration. Chapter 3 Image Registration
Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation
More informationCS231A Midterm Review. Friday 5/6/2016
CS231A Midterm Review Friday 5/6/2016 Outline General Logistics Camera Models Non-perspective cameras Calibration Single View Metrology Epipolar Geometry Structure from Motion Active Stereo and Volumetric
More informationEdge pixel with coordinates (s,t) in S xy has an angle similar to pixel at α(s,t) α(x,y) <A
Image segmentation(10.2.7) SLIDE 1/13 10.2.7 Edge linking and boundary detection Edge detection is always followed by edge linking Local processing AnalyzepixelsinsmallneighbourhoodS xy ofeachedgepoint
More informationTitle: Vanishing Hull: A Geometric Concept for Vanishing Points Detection and Analysis
Pattern Recognition Manuscript Draft Manuscript Number: Title: Vanishing Hull: A Geometric Concept for Vanishing Points Detection and Analysis Article Type: Full Length Article Section/Category: Keywords:
More informationComputer Vision I - Appearance-based Matching and Projective Geometry
Computer Vision I - Appearance-based Matching and Projective Geometry Carsten Rother 01/11/2016 Computer Vision I: Image Formation Process Roadmap for next four lectures Computer Vision I: Image Formation
More informationGraphics Hardware and Display Devices
Graphics Hardware and Display Devices CSE328 Lectures Graphics/Visualization Hardware Many graphics/visualization algorithms can be implemented efficiently and inexpensively in hardware Facilitates interactive
More informationDisplay. Introduction page 67 2D Images page 68. All Orientations page 69 Single Image page 70 3D Images page 71
Display Introduction page 67 2D Images page 68 All Orientations page 69 Single Image page 70 3D Images page 71 Intersecting Sections page 71 Cube Sections page 72 Render page 73 1. Tissue Maps page 77
More informationComplex Numbers, Polar Equations, and Parametric Equations. Copyright 2017, 2013, 2009 Pearson Education, Inc.
8 Complex Numbers, Polar Equations, and Parametric Equations Copyright 2017, 2013, 2009 Pearson Education, Inc. 1 8.5 Polar Equations and Graphs Polar Coordinate System Graphs of Polar Equations Conversion
More informationVectors and the Geometry of Space
Vectors and the Geometry of Space In Figure 11.43, consider the line L through the point P(x 1, y 1, z 1 ) and parallel to the vector. The vector v is a direction vector for the line L, and a, b, and c
More informationData Representation in Visualisation
Data Representation in Visualisation Visualisation Lecture 4 Taku Komura Institute for Perception, Action & Behaviour School of Informatics Taku Komura Data Representation 1 Data Representation We have
More information3D Polygon Rendering. Many applications use rendering of 3D polygons with direct illumination
Rendering Pipeline 3D Polygon Rendering Many applications use rendering of 3D polygons with direct illumination 3D Polygon Rendering What steps are necessary to utilize spatial coherence while drawing
More informationRadiosity. Johns Hopkins Department of Computer Science Course : Rendering Techniques, Professor: Jonathan Cohen
Radiosity Radiosity Concept Global computation of diffuse interreflections among scene objects Diffuse lighting changes fairly slowly across a surface Break surfaces up into some number of patches Assume
More informationAnnouncements. Mosaics. How to do it? Image Mosaics
Announcements Mosaics Project artifact voting Project 2 out today (help session at end of class) http://www.destination36.com/start.htm http://www.vrseattle.com/html/vrview.php?cat_id=&vrs_id=vrs38 Today
More informationCHAPTER 1 Graphics Systems and Models 3
?????? 1 CHAPTER 1 Graphics Systems and Models 3 1.1 Applications of Computer Graphics 4 1.1.1 Display of Information............. 4 1.1.2 Design.................... 5 1.1.3 Simulation and Animation...........
More informationCS Exam 1 Review Problems Fall 2017
CS 45500 Exam 1 Review Problems Fall 2017 1. What is a FrameBuffer data structure? What does it contain? What does it represent? How is it used in a graphics rendering pipeline? 2. What is a Scene data
More informationComputer Graphics 7: Viewing in 3-D
Computer Graphics 7: Viewing in 3-D In today s lecture we are going to have a look at: Transformations in 3-D How do transformations in 3-D work? Contents 3-D homogeneous coordinates and matrix based transformations
More informationImproved Radiance Gradient Computation
Improved Radiance Gradient Computation Jaroslav Křivánek Pascal Gautron Kadi Bouatouch Sumanta Pattanaik Czech Technical University New gradients Gradients by [Křivánek et al. 2005] Figure 1: Right: The
More informationChapter 24. Creating Surfaces for Displaying and Reporting Data
Chapter 24. Creating Surfaces for Displaying and Reporting Data FLUENT allows you to select portions of the domain to be used for visualizing the flow field. The domain portions are called surfaces, and
More informationOrthogonal Projection Matrices. Angel and Shreiner: Interactive Computer Graphics 7E Addison-Wesley 2015
Orthogonal Projection Matrices 1 Objectives Derive the projection matrices used for standard orthogonal projections Introduce oblique projections Introduce projection normalization 2 Normalization Rather
More informationStructure from Motion. Prof. Marco Marcon
Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)
More informationImage-Based Rendering
Image-Based Rendering COS 526, Fall 2016 Thomas Funkhouser Acknowledgments: Dan Aliaga, Marc Levoy, Szymon Rusinkiewicz What is Image-Based Rendering? Definition 1: the use of photographic imagery to overcome
More informationMore Mosaic Madness. CS194: Image Manipulation & Computational Photography. Steve Seitz and Rick Szeliski. Jeffrey Martin (jeffrey-martin.
More Mosaic Madness Jeffrey Martin (jeffrey-martin.com) CS194: Image Manipulation & Computational Photography with a lot of slides stolen from Alexei Efros, UC Berkeley, Fall 2018 Steve Seitz and Rick
More informationGoal. Rendering Complex Scenes on Mobile Terminals or on the web. Rendering on Mobile Terminals. Rendering on Mobile Terminals. Walking through images
Goal Walking though s -------------------------------------------- Kadi Bouatouch IRISA Univesité de Rennes I, Fance Rendeing Comple Scenes on Mobile Teminals o on the web Rendeing on Mobile Teminals Rendeing
More information6.837 Introduction to Computer Graphics Assignment 5: OpenGL and Solid Textures Due Wednesday October 22, 2003 at 11:59pm
6.837 Introduction to Computer Graphics Assignment 5: OpenGL and Solid Textures Due Wednesday October 22, 2003 at 11:59pm In this assignment, you will add an interactive preview of the scene and solid
More informationcalibrated coordinates Linear transformation pixel coordinates
1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial
More informationOcclusion Detection of Real Objects using Contour Based Stereo Matching
Occlusion Detection of Real Objects using Contour Based Stereo Matching Kenichi Hayashi, Hirokazu Kato, Shogo Nishida Graduate School of Engineering Science, Osaka University,1-3 Machikaneyama-cho, Toyonaka,
More information12.7 Tangent Planes and Normal Lines
.7 Tangent Planes and Normal Lines Tangent Plane and Normal Line to a Surface Suppose we have a surface S generated by z f(x,y). We can represent it as f(x,y)-z 0 or F(x,y,z) 0 if we wish. Hence we can
More information5LSH0 Advanced Topics Video & Analysis
1 Multiview 3D video / Outline 2 Advanced Topics Multimedia Video (5LSH0), Module 02 3D Geometry, 3D Multiview Video Coding & Rendering Peter H.N. de With, Sveta Zinger & Y. Morvan ( p.h.n.de.with@tue.nl
More informationComputer Science 426 Midterm 3/11/04, 1:30PM-2:50PM
NAME: Login name: Computer Science 46 Midterm 3//4, :3PM-:5PM This test is 5 questions, of equal weight. Do all of your work on these pages (use the back for scratch space), giving the answer in the space
More informationCHAPTER 3. Single-view Geometry. 1. Consequences of Projection
CHAPTER 3 Single-view Geometry When we open an eye or take a photograph, we see only a flattened, two-dimensional projection of the physical underlying scene. The consequences are numerous and startling.
More informationLesson 6: Contours. 1. Introduction. 2. Image filtering: Convolution. 3. Edge Detection. 4. Contour segmentation
. Introduction Lesson 6: Contours 2. Image filtering: Convolution 3. Edge Detection Gradient detectors: Sobel Canny... Zero crossings: Marr-Hildreth 4. Contour segmentation Local tracking Hough transform
More informationComputer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.
Computer Vision Coordinates Prof. Flávio Cardeal DECOM / CEFET- MG cardeal@decom.cefetmg.br Abstract This lecture discusses world coordinates and homogeneous coordinates, as well as provides an overview
More informationEXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,
School of Computer Science and Communication, KTH Danica Kragic EXAM SOLUTIONS Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, 14.00 19.00 Grade table 0-25 U 26-35 3 36-45
More informationCS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka
CS223b Midterm Exam, Computer Vision Monday February 25th, Winter 2008, Prof. Jana Kosecka Your name email This exam is 8 pages long including cover page. Make sure your exam is not missing any pages.
More informationAnno accademico 2006/2007. Davide Migliore
Robotica Anno accademico 6/7 Davide Migliore migliore@elet.polimi.it Today What is a feature? Some useful information The world of features: Detectors Edges detection Corners/Points detection Descriptors?!?!?
More informationCS 464 Review. Review of Computer Graphics for Final Exam
CS 464 Review Review of Computer Graphics for Final Exam Goal: Draw 3D Scenes on Display Device 3D Scene Abstract Model Framebuffer Matrix of Screen Pixels In Computer Graphics: If it looks right then
More informationFeature Matching and Robust Fitting
Feature Matching and Robust Fitting Computer Vision CS 143, Brown Read Szeliski 4.1 James Hays Acknowledgment: Many slides from Derek Hoiem and Grauman&Leibe 2008 AAAI Tutorial Project 2 questions? This
More informationCamera Model and Calibration
Camera Model and Calibration Lecture-10 Camera Calibration Determine extrinsic and intrinsic parameters of camera Extrinsic 3D location and orientation of camera Intrinsic Focal length The size of the
More informationCamera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration
Camera Calibration Jesus J Caban Schedule! Today:! Camera calibration! Wednesday:! Lecture: Motion & Optical Flow! Monday:! Lecture: Medical Imaging! Final presentations:! Nov 29 th : W. Griffin! Dec 1
More informationStereo Vision. MAN-522 Computer Vision
Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in
More informationTopics in Analytic Geometry Part II
Name Chapter 9 Topics in Analytic Geometry Part II Section 9.4 Parametric Equations Objective: In this lesson you learned how to evaluate sets of parametric equations for given values of the parameter
More informationGame Architecture. 2/19/16: Rasterization
Game Architecture 2/19/16: Rasterization Viewing To render a scene, need to know Where am I and What am I looking at The view transform is the matrix that does this Maps a standard view space into world
More informationRendering. Converting a 3D scene to a 2D image. Camera. Light. Rendering. View Plane
Rendering Pipeline Rendering Converting a 3D scene to a 2D image Rendering Light Camera 3D Model View Plane Rendering Converting a 3D scene to a 2D image Basic rendering tasks: Modeling: creating the world
More informationRay Tracer I: Ray Casting Due date: 12:00pm December 3, 2001
Computer graphics Assignment 5 1 Overview Ray Tracer I: Ray Casting Due date: 12:00pm December 3, 2001 In this assignment you will implement the camera and several primitive objects for a ray tracer. We
More information3D Reconstruction from Scene Knowledge
Multiple-View Reconstruction from Scene Knowledge 3D Reconstruction from Scene Knowledge SYMMETRY & MULTIPLE-VIEW GEOMETRY Fundamental types of symmetry Equivalent views Symmetry based reconstruction MUTIPLE-VIEW
More informationPractical Image and Video Processing Using MATLAB
Practical Image and Video Processing Using MATLAB Chapter 14 Edge detection What will we learn? What is edge detection and why is it so important to computer vision? What are the main edge detection techniques
More informationProject report Augmented reality with ARToolKit
Project report Augmented reality with ARToolKit FMA175 Image Analysis, Project Mathematical Sciences, Lund Institute of Technology Supervisor: Petter Strandmark Fredrik Larsson (dt07fl2@student.lth.se)
More information