Image-Based Lighting. Inserting Synthetic Objects

Save this PDF as:

Size: px
Start display at page:

Transcription

1 Image-Based Lighting : Rendering and Image Processing Alexei Efros with a lot of slides donated by Paul Debevec Inserting Synthetic Objects Why does this look so bad? Wrong camera orientation Wrong lighting No shadows 1

2 Solutions Wrong Camera Orientation Estimate correct camera orientation and renender object Use corresponding points to warp the object/scene Only works for small warps and/or mostly planar objects Lighting & Shadows Estimate (eyeball) all the light sources in the scene and simulate it in your virtual rendering Now can use shadow matting to put in shadows But what happens if lighting is complex? Extended light sources, mutual illumination, etc. Environment Maps Simple solution for shiny objects Models complex lighting as a panoramic image i.e. amount of radiance coming in from each direction 2

3 Environment Mapping Reflected ray: r=2(n v)n-v projector function converts reflection vector (x, y, z) to texture image (u, v) viewer n v r reflective surface environment texture image Texture is transferred in the direction of the reflected ray from the environment map onto the object What is in the map? What approximations are made? The map should contain a view of the world with the point of interest on the object as the eye We can t store a separate map for each point, so one map is used with the eye at the center of the object Introduces distortions in the reflection, but the eye doesn t notice Distortions are minimized for a small object in a large room The object will not reflect itself! 3

4 Environment Maps The environment map may take one of several forms: Cubic mapping Spherical mapping other Describes the shape of the surface on which the map resides Determines how the map is generated and how it is indexed Cubic Mapping The map resides on the surfaces of a cube around the object Typically, align the faces of the cube with the coordinate axes To generate the map: For each face of the cube, render the world from the center of the object with the cube face as the image plane Rendering can be arbitrarily complex (it s off-line) To use the map: Index the R ray into the correct cube face Compute texture coordinates 4

5 Cubic Map Example Sphere Mapping Map lives on a sphere To generate the map: Render a spherical panorama from the designed center point To use the map: Use the orientation of the R ray to index directly into the sphere 5

6 Example What about real scenes? from Terminator 2 6

7 Real environment maps We can use photographs to capture environment maps The first use of panoramic mosaics How do we deal with light sources? Sun, lights, etc? They are much much brighter than the rest of the enviarnment User High Dynamic Range photography, of course! Several ways to acquire environment maps: Stitching mosaics Fisheye lens Mirrored Balls Stitching HDR mosaics 7

8 Scanning Panoramic Cameras Pros: very high res (10K x 7K+) Full sphere in one scan no stitching Good dynamic range, some are HDR Issues: More expensive Scans take a while Companies: Panoscan, Sphereon See also 8

9 Fisheye Images 9

10 Mirrored Sphere 10

11 Sources of Mirrored Balls 2-inch chrome balls ~ \$20 ea. McMaster-Carr Supply Company inch large gazing balls Baker s Lawn Ornaments Hollow Spheres, 2in 4in Dube Juggling Equipment FAQ on 11

12 0.34 => 59% Reflective Calibrating Mirrored Sphere Reflectivity 0.58 Real-World HDR Lighting Environments Funston Beach Eucalyptus Grove Uffizi Gallery Grace Cathedral Lighting Environments from the Light Probe Image Gallery: 12

13 Acquiring the Light Probe Assembling the Light Probe 13

14 Not just shiny We have captured a true radiance map We can treat it as an extended (e.g spherical) light source Can use Global Illumination to simulate light transport in the scene So, all objects (not just shiny) can be lighted What s the limitation? Illumination Results 14

15 Comparison: Radiance map versus single image Putting it all together Synthetic Objects + Real light! 15

16 CG Objects Illuminated by a Traditional CG Light Source Illuminating Objects using Measurements of Real Light Object Light Environment assigned glow material property in Greg Ward s RADIANCE system. 16

17 Paul Debevec. A Tutorial on Image-Based Lighting. IEEE Computer Graphics and Applications, Jan/Feb Rendering with Natural Light SIGGRAPH 98 Electronic Theater 17

18 RNL Environment mapped onto interior of large cube MOVIE! 18

Image-Based Lighting. Eirik Holmøyvik. with a lot of slides donated by Paul Debevec

Image-Based Lighting Eirik Holmøyvik with a lot of slides donated by Paul Debevec 15-463: Computational Photography Alexei Efros, CMU, Fall 2006 Inserting Synthetic Objects Why does this look so bad? Wrong

Image-Based Lighting

Image-Based Lighting Eirik Holmøyvik CS194: Image Manipulation & Computational Photography with a lot of slides Alexei Efros, UC Berkeley, Fall 2014 donated by Paul Debevec Inserting Synthetic Objects

Image-Based Lighting : Computational Photography Alexei Efros, CMU, Fall Eirik Holmøyvik. with a lot of slides donated by Paul Debevec

Image-Based Lighting Eirik Holmøyvik with a lot of slides donated by Paul Debevec 15-463: Computational Photography Alexei Efros, CMU, Fall 2011 Inserting Synthetic Objects Why does this look so bad? Wrong

The 7d plenoptic function, indexing all light.

Previous Lecture The 7d plenoptic function, indexing all light. Lightfields: a 4d (not 5d!) data structure which captures all outgoing light from a region and permits reconstruction of arbitrary synthetic

Image-based Lighting (Part 2)

Image-based Lighting (Part 2) 10/19/17 T2 Computational Photography Derek Hoiem, University of Illinois Many slides from Debevec, some from Efros, Kevin Karsch Today Brief review of last class Show how

Image-based Lighting

Image-based Lighting 10/17/15 T2 Computational Photography Derek Hoiem, University of Illinois Many slides from Debevec, some from Efros Next week Derek away for ICCV (Venice) Zhizhong and Aditya will

Rendering Synthetic Objects into Real Scenes. based on [Debevec98]

Rendering Synthetic Objects into Real Scenes based on [Debevec98] Compositing of synthetic objects Geometry consistency needed: geometric model of synthetic objects needed: (coarse) geometric model of

CS 354R: Computer Game Technology

CS 354R: Computer Game Technology Texture and Environment Maps Fall 2018 Texture Mapping Problem: colors, normals, etc. are only specified at vertices How do we add detail between vertices without incurring

CS4670/5760: Computer Vision

CS4670/5760: Computer Vision Kavita Bala! Lecture 28: Photometric Stereo Thanks to ScoC Wehrwein Announcements PA 3 due at 1pm on Monday PA 4 out on Monday HW 2 out on weekend Next week: MVS, sfm Last

Computer Graphics. Lecture 9 Environment mapping, Mirroring

Computer Graphics Lecture 9 Environment mapping, Mirroring Today Environment Mapping Introduction Cubic mapping Sphere mapping refractive mapping Mirroring Introduction reflection first stencil buffer

High Dynamic Range Lighting Paul Debevec, USC Institute for Creative Technologies. March 24, Game Developer s Conference 1

March 4, 004 High Dynamic Range Lighting Paul Debevec University of Southern California Institute for Creative Technologies March 4, 004 5:30 6:30 pm www.debevec.org/ibl004/ Scenes lit with point light

Computer Graphics for Large- Scale Immersive Theaters

Computer Graphics for Large- Scale Immersive Theaters Immersive Rendering Basics Brad Thompson Lead Animator Spitz, Inc. bthompson@spitzinc.com Aesthetic Considerations: Understanding the space Completely

High Dynamic Range Image Texture Mapping based on VRML

High Dynamic Range Image Texture Mapping based on VRML Sung-Ye Kim and Byoung-Tae Choi 3D Graphics Research Team, Virtual Reality Research and Development Department, Computer Software Research Laboratory,

Introduction to Game FX (1/2) Improve the visual & sound game effects Includes : Combat FX Environment FX Character FX Scene FX Sound FX Post-processi

Game FX 1 Introduction to Game FX (1/2) Improve the visual & sound game effects Includes : Combat FX Environment FX Character FX Scene FX Sound FX Post-processing after rendering FX editor needed General

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac

Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Stitching = alignment + blending geometrical registration

Texture. Texture Maps

Texture Texture maps! Surface color and transparency! Environment and irradiance maps! Reflectance maps! Shadow maps! Displacement and bump maps Level of detail hierarchy Procedural shading and texturing

Announcements. Mosaics. Image Mosaics. How to do it? Basic Procedure Take a sequence of images from the same position =

Announcements Project 2 out today panorama signup help session at end of class Today mosaic recap blending Mosaics Full screen panoramas (cubic): http://www.panoramas.dk/ Mars: http://www.panoramas.dk/fullscreen3/f2_mars97.html

Texture. Detail Representation

Page 1 Texture Procedural shading and texturing Applied and projected textures Material / light properties Shadow maps Spherical and higher order textures Spherical mappings Environment and irradiance

Mosaics VR Seattle: http://www.vrseattle.com/ Full screen panoramas (cubic): http://www.panoramas.dk/ Mars: http://www.panoramas.dk/fullscreen3/f2_mars97.html Today s Readings Szeliski and Shum paper (sections

Overview. Hierarchy. Level of detail hierarchy Texture maps Procedural shading and texturing Texture synthesis and noise.

Overview Level of detail hierarchy Texture maps Procedural shading and texturing Texture synthesis and noise Hierarchy Physics Computer Graphics Geometrical optics Macro-structures Transport Micro-structures

Multiple Importance Sampling

Multiple Importance Sampling Multiple Importance Sampling Reflection of a circular light source by a rough surface Radius Shininess Sampling the light source f()() xgxdx Sampling the BRDF Page 1 Multiple

Consider a partially transparent object that is illuminated with two lights, one visible from each side of the object. Start with a ray from the eye

Ray Tracing What was the rendering equation? Motivate & list the terms. Relate the rendering equation to forward ray tracing. Why is forward ray tracing not good for image formation? What is the difference

A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction

A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction Jaemin Lee and Ergun Akleman Visualization Sciences Program Texas A&M University Abstract In this paper we present a practical

Light Transport CS434. Daniel G. Aliaga Department of Computer Science Purdue University

Light Transport CS434 Daniel G. Aliaga Department of Computer Science Purdue University Topics Local and Global Illumination Models Helmholtz Reciprocity Dual Photography/Light Transport (in Real-World)

Omni-directional stereoscopy

Omni-directional stereoscopy Paul Bourke (WASP, UWA) Motivation Correct stereoscopic views require a matching relationship between viewing geometry and rendering geometry. That is, viewer position/relationship

Reflection Mapping

Image-Based Lighting A Photometric Approach to Rendering and Compositing Paul Debevec Computer Science Division University of California at Berkeley http://www.cs.berkeley.edu/~debevec August 1999 Reflection

Real-Time Image Based Lighting in Software Using HDR Panoramas

Real-Time Image Based Lighting in Software Using HDR Panoramas Jonas Unger, Magnus Wrenninge, Filip Wänström and Mark Ollila Norrköping Visualization and Interaction Studio Linköping University, Sweden

Image Warping and Mosacing

Image Warping and Mosacing 15-463: Rendering and Image Processing Alexei Efros with a lot of slides stolen from Steve Seitz and Rick Szeliski Today Mosacs Image Warping Homographies Programming Assignment

CS294-13: Special Topics Lecture #8 Advanced Computer Graphics University of California, Berkeley Monday, 28 September 2009 Shadow and Environment Maps Lecture #8: Monday, 28 September 2009 Lecturer: Ravi

BRYCE 5 Mini Tutorial

BRYCE 5 Mini Tutorial Abstract Panoramics by David Brinnen March 2005 HTML Version by Hans-Rudolf Wernli Blotchy Splashes of Colour > If you have photographed a panorama with your camera once, you know

CIS 536/636 Introduction to Computer Graphics. Kansas State University. CIS 536/636 Introduction to Computer Graphics

2 Lecture Outline Surface Detail 3 of 5: Mappings OpenGL Textures William H. Hsu Department of Computing and Information Sciences, KSU KSOL course pages: http://bit.ly/hgvxlh / http://bit.ly/evizre Public

CMSC427 Advanced shading getting global illumination by local methods. Credit: slides Prof. Zwicker

CMSC427 Advanced shading getting global illumination by local methods Credit: slides Prof. Zwicker Topics Shadows Environment maps Reflection mapping Irradiance environment maps Ambient occlusion Reflection

IMAGE BASED RENDERING: Using High Dynamic Range Photographs to Light Architectural Scenes

IMAGE BASED RENDERING 1 IMAGE BASED RENDERING: Using High Dynamic Range Photographs to Light Architectural Scenes KATHLEEN CHENEY MEHLIKA INANICI University of Washington Abstract Image Based Rendering

CS6670: Computer Vision

CS6670: Computer Vision Noah Snavely Lecture 7: Image Alignment and Panoramas What s inside your fridge? http://www.cs.washington.edu/education/courses/cse590ss/01wi/ Projection matrix intrinsics projection

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier

Computer Vision 2 SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung Computer Vision 2 Dr. Benjamin Guthier 3. HIGH DYNAMIC RANGE Computer Vision 2 Dr. Benjamin Guthier Pixel Value Content of this

Projective Texture Mapping with Full Panorama

EUROGRAPHICS 2002 / G. Drettakis and H.-P. Seidel Volume 21 (2002), Number 3 (Guest Editors) Projective Texture Mapping with Full Panorama Dongho Kim and James K. Hahn Department of Computer Science, The

Computer Graphics Lecture 11

1 / 14 Computer Graphics Lecture 11 Dr. Marc Eduard Frîncu West University of Timisoara May 15th 2012 2 / 14 Outline 1 Introduction 2 Transparency 3 Reflection 4 Recap 3 / 14 Introduction light = local

Image Based Lighting with Near Light Sources

Image Based Lighting with Near Light Sources Shiho Furuya, Takayuki Itoh Graduate School of Humanitics and Sciences, Ochanomizu University E-mail: {shiho, itot}@itolab.is.ocha.ac.jp Abstract Recent some

Image Based Lighting with Near Light Sources

Image Based Lighting with Near Light Sources Shiho Furuya, Takayuki Itoh Graduate School of Humanitics and Sciences, Ochanomizu University E-mail: {shiho, itot}@itolab.is.ocha.ac.jp Abstract Recent some

Research White Paper WHP 143. Multi-camera radiometric surface modelling for image-based re-lighting BRITISH BROADCASTING CORPORATION.

Research White Paper WHP 143 11 January 2007 Multi-camera radiometric surface modelling for image-based re-lighting Oliver Grau BRITISH BROADCASTING CORPORATION Multi-camera radiometric surface modelling

CS 498 VR. Lecture 20-4/11/18. go.illinois.edu/vrlect20

CS 498 VR Lecture 20-4/11/18 go.illinois.edu/vrlect20 Review from last lecture Texture, Normal mapping Three types of optical distortion? How does texture mipmapping work? Improving Latency and Frame Rates

Introduction to 3D Concepts

PART I Introduction to 3D Concepts Chapter 1 Scene... 3 Chapter 2 Rendering: OpenGL (OGL) and Adobe Ray Tracer (ART)...19 1 CHAPTER 1 Scene s0010 1.1. The 3D Scene p0010 A typical 3D scene has several

CS635 Spring Department of Computer Science Purdue University

Light Transport CS635 Spring 2010 Daniel G Aliaga Daniel G. Aliaga Department of Computer Science Purdue University Topics Local and GlobalIllumination Models Helmholtz Reciprocity Dual Photography/Light

Modeling Light. Michal Havlik : Computational Photography Alexei Efros, CMU, Fall 2007

Modeling Light Michal Havlik 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 The Plenoptic Function Figure by Leonard McMillan Q: What is the set of all things that we can ever see? A: The

Image-Based Modeling and Rendering

Image-Based Modeling and Rendering Richard Szeliski Microsoft Research IPAM Graduate Summer School: Computer Vision July 26, 2013 How far have we come? Light Fields / Lumigraph - 1996 Richard Szeliski

Representing and Computing Polarized Light in a Ray Tracer

Representing and Computing Polarized Light in a Ray Tracer A Technical Report in STS 4600 Presented to the Faculty of the School of Engineering and Applied Science University of Virginia in Partial Fulfillment

Recap of Previous Lecture

Recap of Previous Lecture Matting foreground from background Using a single known background (and a constrained foreground) Using two known backgrounds Using lots of backgrounds to capture reflection and

Image stitching. Digital Visual Effects Yung-Yu Chuang. with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac

Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac Image stitching Stitching = alignment + blending geometrical registration

The Shading Probe: Fast Appearance Acquisition for Mobile AR

The Shading Probe: Fast Appearance Acquisition for Mobile AR Dan Andrei Calian * Kenny Mitchell Derek Nowrouzezahrai Jan Kautz * * University College London Disney Research Zürich University of Montreal

Scene Modeling for a Single View

Scene Modeling for a Single View René MAGRITTE Portrait d'edward James with a lot of slides stolen from Steve Seitz and David Brogan, 15-463: Computational Photography Alexei Efros, CMU, Fall 2005 Classes

Image stitching. Announcements. Outline. Image stitching

Announcements Image stitching Project #1 was due yesterday. Project #2 handout will be available on the web later tomorrow. I will set up a webpage for artifact voting soon. Digital Visual Effects, Spring

Scene Modeling for a Single View

Scene Modeling for a Single View René MAGRITTE Portrait d'edward James with a lot of slides stolen from Steve Seitz and David Brogan, 15-463: Computational Photography Alexei Efros, CMU, Spring 2010 Breaking

Recursion and Data Structures in Computer Graphics. Ray Tracing

Recursion and Data Structures in Computer Graphics Ray Tracing 1 Forward Ray Tracing imagine that you take a picture of a room using a camera exactly what is the camera sensing? light reflected from the

lecture 18 - ray tracing - environment mapping - refraction

lecture 18 - ray tracing - environment mapping - refraction Recall Ray Casting (lectures 7, 8) for each pixel (x,y) { cast a ray through that pixel into the scene, and find the closest surface along the

Digitization of 3D Objects for Virtual Museum

Digitization of 3D Objects for Virtual Museum Yi-Ping Hung 1, 2 and Chu-Song Chen 2 1 Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan 2 Institute of

BioTechnology. An Indian Journal FULL PAPER. Trade Science Inc. A wavelet based real-time rendering technology for indoor mixed reality ABSTRACT

[Type text] [Type text] [Type text] ISSN : 0974-7435 Volume 10 Issue 24 2014 BioTechnology An Indian Journal FULL PAPER BTAIJ, 10(24), 2014 [15095-15100] A wavelet based real-time rendering technology

Image-based modeling (IBM) and image-based rendering (IBR)

Image-based modeling (IBM) and image-based rendering (IBR) CS 248 - Introduction to Computer Graphics Autumn quarter, 2005 Slides for December 8 lecture The graphics pipeline modeling animation rendering

Scene Modeling for a Single View

Scene Modeling for a Single View René MAGRITTE Portrait d'edward James CS194: Image Manipulation & Computational Photography with a lot of slides stolen from Alexei Efros, UC Berkeley, Fall 2014 Steve

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 12 130228 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Panoramas, Mosaics, Stitching Two View Geometry

Announcements. Written Assignment 2 is out see the web page. Computer Graphics

Announcements Written Assignment 2 is out see the web page 1 Texture and other Mappings Shadows Texture Mapping Bump Mapping Displacement Mapping Environment Mapping Watt Chapter 8 COMPUTER GRAPHICS 15-462

Today s lecture. Image Alignment and Stitching. Readings. Motion models

Today s lecture Image Alignment and Stitching Computer Vision CSE576, Spring 2005 Richard Szeliski Image alignment and stitching motion models cylindrical and spherical warping point-based alignment global

More Single View Geometry

More Single View Geometry with a lot of slides stolen from Steve Seitz Cyclops Odilon Redon 1904 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Final Projects Are coming up fast! Undergrads

More Single View Geometry

More Single View Geometry with a lot of slides stolen from Steve Seitz Cyclops Odilon Redon 1904 15-463: Computational Photography Alexei Efros, CMU, Fall 2006 Final Projects Are coming up fast! Undergrads

COMP30019 Graphics and Interaction Perspective & Polygonal Geometry

COMP30019 Graphics and Interaction Perspective & Polygonal Geometry Department of Computing and Information Systems The Lecture outline Introduction Perspective Geometry Virtual camera Centre of projection

Immersive Rendering Basics and Aesthetics

Course #25 Computer Graphics for Large Scale Immersive Theaters Immersive Rendering Basics and Aesthetics Brad Thompson Lead Animator Spitz, Inc. Bthompson@spitzinc.com Introducing the Frame This is not

Image warping and stitching

Image warping and stitching May 4 th, 2017 Yong Jae Lee UC Davis Last time Interactive segmentation Feature-based alignment 2D transformations Affine fit RANSAC 2 Alignment problem In alignment, we will

CSE528 Computer Graphics: Theory, Algorithms, and Applications

CSE528 Computer Graphics: Theory, Algorithms, and Applications Hong Qin State University of New York at Stony Brook (Stony Brook University) Stony Brook, New York 11794--4400 Tel: (631)632-8450; Fax: (631)632-8334

HDR image based lighting: from acquisition to render. Artem Krizhanovskiy Technical Artist Wargaming

HDR image based lighting: from acquisition to render Artem Krizhanovskiy Technical Artist Wargaming Image based Lighting: applications PBR : Maxwell, Corona, V-ray, etc. Content creation software: SP,

More Mosaic Madness. CS194: Image Manipulation & Computational Photography. Steve Seitz and Rick Szeliski. Jeffrey Martin (jeffrey-martin.

More Mosaic Madness Jeffrey Martin (jeffrey-martin.com) CS194: Image Manipulation & Computational Photography with a lot of slides stolen from Alexei Efros, UC Berkeley, Fall 2018 Steve Seitz and Rick

Introduction to Image Processing and Computer Vision. -- Panoramas and Blending --

Introduction to Image Processing and Computer Vision -- Panoramas and Blending -- Winter 2013/14 Ivo Ihrke Panoramas Mosaics and Panoramas - Outline - Perspective Panoramas - Hardware-Based - Software-Based

6.098 Digital and Computational Photography Advanced Computational Photography. Panoramas. Bill Freeman Frédo Durand MIT - EECS

6.098 Digital and Computational Photography 6.882 Advanced Computational Photography Panoramas Bill Freeman Frédo Durand MIT - EECS Lots of slides stolen from Alyosha Efros, who stole them from Steve Seitz

Assignment 2 Ray Tracing

Assignment 2 Ray Tracing Overview The concept of ray tracing: a technique for generating an image by tracing the path of light through pixels in an image plane and simulating the effects of its encounters

Panoramas. Why Mosaic? Why Mosaic? Mosaics: stitching images together. Why Mosaic? Olivier Gondry. Bill Freeman Frédo Durand MIT - EECS

Olivier Gondry 6.098 Digital and Computational Photography 6.882 Advanced Computational Photography Panoramas Director of music video and commercial Special effect specialist (Morphing, rotoscoping) Today

Image warping and stitching

Image warping and stitching May 5 th, 2015 Yong Jae Lee UC Davis PS2 due next Friday Announcements 2 Last time Interactive segmentation Feature-based alignment 2D transformations Affine fit RANSAC 3 Alignment

Projection simulator to support design development of spherical immersive display

Projection simulator to support design development of spherical immersive display Wataru Hashimoto, Yasuharu Mizutani, and Satoshi Nishiguchi Faculty of Information Sciences and Technology, Osaka Institute

Faces and Image-Based Lighting

Announcements Faces and Image-Based Lighting Project #3 artifacts voting Final project: Demo on 6/25 (Wednesday) 13:30pm in this room Reports and videos due on 6/26 (Thursday) 11:59pm Digital Visual Effects,

3D graphics, raster and colors CS312 Fall 2010

Computer Graphics 3D graphics, raster and colors CS312 Fall 2010 Shift in CG Application Markets 1989-2000 2000 1989 3D Graphics Object description 3D graphics model Visualization 2D projection that simulates

We want to put a CG object in this room

The Problem Temporally varying lighting What if we rearrange the furniture frequently? Mirror surfaces Diffuse surfaces We want to put a CG object in this room Are we allowed to put light probes on the

Multiple View Geometry

Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V

Homographies and RANSAC

Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position

COMP 175: Computer Graphics April 11, 2018

Lecture n+1: Recursive Ray Tracer2: Advanced Techniques and Data Structures COMP 175: Computer Graphics April 11, 2018 1/49 Review } Ray Intersect (Assignment 4): questions / comments? } Review of Recursive

Scalable multi-gpu cloud raytracing with OpenGL

Scalable multi-gpu cloud raytracing with OpenGL University of Žilina Digital technologies 2014, Žilina, Slovakia Overview Goals Rendering distant details in visualizations Raytracing Multi-GPU programming

Recollection. Models Pixels. Model transformation Viewport transformation Clipping Rasterization Texturing + Lights & shadows

Recollection Models Pixels Model transformation Viewport transformation Clipping Rasterization Texturing + Lights & shadows Can be computed in different stages 1 So far we came to Geometry model 3 Surface

1-(800) 877-2745 www.ashlar-vellum.com Textures Tutorial Using Cobalt, Xenon, Argon Copyright 2008-2014 Vellum Investment Partners dba Ashlar-Vellum. All rights reserved. Ashlar-Vellum Cobalt, Xenon &

Topic 12: Texture Mapping. Motivation Sources of texture Texture coordinates Bump mapping, mip-mapping & env mapping

Topic 12: Texture Mapping Motivation Sources of texture Texture coordinates Bump mapping, mip-mapping & env mapping Texture sources: Photographs Texture sources: Procedural Texture sources: Solid textures

Topic 11: Texture Mapping 11/13/2017. Texture sources: Solid textures. Texture sources: Synthesized

Topic 11: Texture Mapping Motivation Sources of texture Texture coordinates Bump mapping, mip mapping & env mapping Texture sources: Photographs Texture sources: Procedural Texture sources: Solid textures

Turn your movie file into the homework folder on the server called Lights, Camera, Action.

CS32 W11 Homework 3: Due MONDAY, APRIL 18 Now let s put the ball in a world of your making and have some fun. Create a simple AND WE MEAN SIMPLE environment for one of your ball bounces. You will assign

Computed Photography - Final Project Endoscope Exploration on Knee Surface

15-862 Computed Photography - Final Project Endoscope Exploration on Knee Surface Chenyu Wu Robotics Institute, Nov. 2005 Abstract Endoscope is widely used in the minimally invasive surgery. However the

o Basic signal processing o Filtering, resampling, warping,... Rendering o Polygon rendering pipeline o Ray tracing Modeling

Background COS526: Advanced Computer Graphics Tom Funkhouser Fall 2010 Image Processing o Basic signal processing o Filtering, resampling, warping,... Rendering o Polygon rendering pipeline o Ray tracing

Computergrafik. Matthias Zwicker Universität Bern Herbst 2016

Computergrafik Matthias Zwicker Universität Bern Herbst 2016 Today More shading Environment maps Reflection mapping Irradiance environment maps Ambient occlusion Reflection and refraction Toon shading

Acquisition and Visualization of Colored 3D Objects

Acquisition and Visualization of Colored 3D Objects Kari Pulli Stanford University Stanford, CA, U.S.A kapu@cs.stanford.edu Habib Abi-Rached, Tom Duchamp, Linda G. Shapiro and Werner Stuetzle University

Computer Graphics. - Texturing Methods -

Computer Graphics - Texturing Methods - Overview Last time BRDFs Shading Today Texturing Texture parameterization Procedural methods Procedural textures Fractal landscapes Next lecture Texture filtering

Recall: Indexing into Cube Map

Recall: Indexing into Cube Map Compute R = 2(N V)N-V Object at origin Use largest magnitude component of R to determine face of cube Other 2 components give texture coordinates V R Cube Map Layout Example

More and More on Light Fields. Last Lecture

More and More on Light Fields Topics in Image-Based Modeling and Rendering CSE291 J00 Lecture 4 Last Lecture Re-review with emphasis on radiometry Mosaics & Quicktime VR The Plenoptic function The main

Topic 11: Texture Mapping 10/21/2015. Photographs. Solid textures. Procedural

Topic 11: Texture Mapping Motivation Sources of texture Texture coordinates Bump mapping, mip mapping & env mapping Topic 11: Photographs Texture Mapping Motivation Sources of texture Texture coordinates

CS4670: Computer Vision

CS467: Computer Vision Noah Snavely Lecture 13: Projection, Part 2 Perspective study of a vase by Paolo Uccello Szeliski 2.1.3-2.1.6 Reading Announcements Project 2a due Friday, 8:59pm Project 2b out Friday

03 RENDERING PART TWO

03 RENDERING PART TWO WHAT WE HAVE SO FAR: GEOMETRY AFTER TRANSFORMATION AND SOME BASIC CLIPPING / CULLING TEXTURES AND MAPPING MATERIAL VISUALLY DISTINGUISHES 2 OBJECTS WITH IDENTICAL GEOMETRY FOR NOW,

Image-Based Modeling and Rendering. Image-Based Modeling and Rendering. Final projects IBMR. What we have learnt so far. What IBMR is about

Image-Based Modeling and Rendering Image-Based Modeling and Rendering MIT EECS 6.837 Frédo Durand and Seth Teller 1 Some slides courtesy of Leonard McMillan, Wojciech Matusik, Byong Mok Oh, Max Chen 2