Computational Cameras: Exploiting Spatial- Angular Temporal Tradeoffs in Photography
|
|
- Jessie Jacobs
- 5 years ago
- Views:
Transcription
1 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras: Exploiting Spatial- Angular Temporal Tradeoffs in Photography Amit Agrawal Mitsubishi Electric Research Labs (MERL) Cambridge, MA, USA
2 Mitsubishi Electric Research Labs (MERL) Computational Cameras Where are the cameras?
3 Mitsubishi Electric Research Labs (MERL) Computational Cameras Cameras in Mobile Phones Source: isuppli
4 Mitsubishi Electric Research Labs (MERL) Computational Cameras Have Cameras Evolved? Lens Based Camera Obscura, 1568 Digital Cameras
5 Mitsubishi Electric Research Labs (MERL) Computational Cameras Conventional Cameras Tradeoffs in photography Aperture size, shutter speed, ISO Fast lens More light but low depth of field Allows small shutter time Macro, Wildlife, Sports High ISO Low light scenes, but more noise
6 Mitsubishi Electric Research Labs (MERL) Computational Cameras Have Projectors Evolved? Similar trends in form factor/cost Film/Slide projectors Digital projectors Pocket Projectors Pico Projectors Projectors in smartphones
7 Mitsubishi Electric Research Labs (MERL) Computational Cameras Projector vs Cameras Current projectors offer capabilities far beyond current cameras Each projector pixel can be independently controlled Allows coding and modulation of outgoing light How about cameras where each pixel can be independently controlled? Allow coding and modulation of incoming light?
8 Mitsubishi Electric Research Labs (MERL) Computational Cameras Projectors vs Cameras Exposure, Frame Rate, Resolution etc. High level controls Brightness, color temperature Per Pixel Control?
9 Mitsubishi Electric Research Labs (MERL) Computational Cameras Projectors vs Cameras Exposure, Frame Rate, Resolution etc. High level controls Brightness, color temperature Computational Cameras Per Pixel Control?
10 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras Flutter Shutter Camera Coded Aperture Mask based light field camera Reinterpretable Camera Wide Angle light field camera
11 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras Flutter Shutter Camera Coded Aperture Mask based light field camera Reinterpretable Camera Wide Angle light field camera
12 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras Camera Coding/Modulation Dimension Flutter Shutter Time (Exposure) Coded Aperture Space Light Field Camera Space and Angle Reinterpretable Camera Space, Time, Angle Flexible Voxels Space and Time
13 Mitsubishi Electric Research Labs (MERL) Computational Cameras
14 Mitsubishi Electric Research Labs (MERL) Computational Cameras Coded Exposure [Raskar, Agrawal, Tumblin SIGGRAPH 2006]
15 Mitsubishi Electric Research Labs (MERL) Computational Cameras Coded Exposure (Flutter Shutter) Camera Raskar, Agrawal, Tumblin [Siggraph2006] Coding in Time: Shutter is opened and closed
16 Mitsubishi Electric Research Labs (MERL) Computational Cameras Blurring == Convolution Sharp Photo Blurred Photo PSF == Sinc Function Traditional Camera: Shutter is OPEN: Box Filter ω
17 Mitsubishi Electric Research Labs (MERL) Computational Cameras Sharp Photo Blurred Photo PSF == Broadband Function Preserves High Spatial Frequencies Flutter Shutter: Shutter is OPEN and CLOSED
18 Mitsubishi Electric Research Labs (MERL) Traditional Computational Cameras Coded Exposure Deblurred Image Deblurred Image Image of Static Object
19
20 Coded Exposure (Flutter Shutter) Camera Raskar, Agrawal, Tumblin [Siggraph2006] Coding in Time: Shutter is opened and closed
21 Mitsubishi Electric Research Labs (MERL) Computational Cameras Flutter Shutter Video Camera Pointgrey Dragonfly2 Camera Use Trigger Mode 5 On-chip, Additional Cost = $0
22 Mitsubishi Electric Research Labs (MERL) How to handle focus blur? Computational Cameras
23 Mitsubishi Electric Research Labs (MERL) Coded Exposure (Flutter Shutter) Raskar, Agrawal, Tumblin SIGGRAPH 2006 Computational Cameras Coded Aperture with Veeraraghavan, Raskar, Tumblin, & Mohan, SIGGRAPH 2007 Temporal 1-D broadband code: Motion Deblurring Spatial 2-D broadband code: Focus Deblurring
24 Mitsubishi Electric Research Labs (MERL) Computational Cameras LED In Focus Photo
25 Mitsubishi Electric Research Labs (MERL) Computational Cameras Out of Focus Photo: Open Aperture
26 Mitsubishi Electric Research Labs (MERL) Computational Cameras Out of Focus Photo: Coded Aperture
27 Blurred Photos Open Aperture Coded Aperture, 7 * 7 Mask
28 Deblurred Photos Open Aperture Coded Aperture, 7 * 7 Mask
29 Mitsubishi Electric Research Labs (MERL) Captured Blurred Photo Computational Cameras
30 Mitsubishi Electric Research Labs (MERL) Refocused on Person Computational Cameras
31
32 Mitsubishi Electric Research Labs (MERL) Computational Cameras Coded Imaging Blocking Light == More Information Coded Exposure Coding in Time Coded Aperture Coding in Space
33 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras Camera Coding/Modulation Dimension Flutter Shutter Time (Exposure) Coded Aperture Space Light Field Camera Space and Angle
34 Mask? Mask Sensor Mask Sensor Full Resolution Digital Refocusing: Coded Aperture Camera 4D Light Field from 2D Photo: Heterodyne Light Field Camera
35 Mitsubishi Electric Research Labs (MERL) Computational Cameras Lytro: Lenslet-based Light Field camera Adelson and Wang, 1992, Ng et al. 2005
36 Mask based Light Field Camera (SIGGRAPH 2007) Sensor Mask Sum of Cosines Mask Pinhole Array Mask Tiled Broadband Mask
37 MERL, Northwestern Univ. Mask-Enhanced Cameras: Heterodyned Light Fields & Coded Aperture Veeraraghavan, Raskar, Agrawal, Mohan & Tumblin Optical Heterodyning High Freq Carrier MHz Receiver: Demodulation Baseband Audio Signal Incoming Signal Reference Carrier Main Lens Object Mask Sensor Software Demodulation Recovered Light Field Photographic Signal (Light Field) Carrier Incident Modulated Signal Reference Carrier
38 Captured Light Field Digital Refocusing
39 Recovering Full Resolution 2D Image For in-focus scene Inserting Mask == Spatially Varying Image Attenuation Compensate using calibration image Full Resolution Image In Focus Out of Focus Captured Photo In Focus Out of Focus Calibration Photo of Pinhole Array
40 Recovered Image In Focus Out of Focus
41 Lens Glare Reduction using Light Field
42 Mitsubishi Electric Research Labs (MERL) Computational Cameras Effects of Glare on Image Hard to model, Low Frequency in 2D But reflection glare is outlier in 4D ray-space Sensor b a Lens Inter-reflections Angular Variation at pixel a
43 Mitsubishi Electric Research Labs (MERL) Computational Cameras Captured Photo: LED On
44 Mitsubishi Electric Research Labs (MERL) v Computational Cameras u y x
45 Mitsubishi Electric Research Labs (MERL) Computational Cameras Sequence of Sub-Aperture Views Traditional Camera Photo Glare Reduced Photo
46 Mitsubishi Electric Research Labs (MERL) Computational Cameras
47 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras Camera Coding/Modulation Dimension Flutter Shutter Time (Exposure) Coded Aperture Space Light Field Camera Space and Angle Reinterpretable Camera Space, Time, Angle
48 Captured Photo
49 Video from Single-Shot (Temporal Frames)
50 Captured Photo
51 Rotating Doll in Focus
52
53
54
55
56
57
58
59
60
61 Reinterpretable Camera Resolution tradeoff for Conventional Imaging Fixed before capture video camera, lightfield camera Scene independent Resolution tradeoff for Reinterpretable Camera Variable in post-capture Scene dependent Different for different parts of the scene/captured photo
62 Captured 2D Photo
63 Captured 2D Photo Static Scene Parts In-Focus High Resolution 2D Image
64 Captured 2D Photo Static Scene Parts In-Focus Out of Focus High Resolution 2D Image 4D Light Field
65 Captured 2D Photo Static Scene Parts Dynamic Scene Parts In-Focus Out of Focus In-Focus High Resolution 2D Image 4D Light Field Video
66 Captured 2D Photo Static Scene Parts Dynamic Scene Parts In-Focus Out of Focus In-Focus Out of Focus High Resolution 2D Image 4D Light Field Video 1D Parallax + Motion
67 Coded Aperture Optical Heterodyning Reinterpretable Imager Static Aperture Mask Sensor Static Mask Sensor Dynamic Aperture Mask Static Mask Sensor SIGGRAPH 2007 Veeraraghavan et al. SIGGRAPH 2007 This Paper Digital Refocusing
68 Coded Aperture Optical Heterodyning Reinterpretable Imager Static Aperture Mask Sensor Static Mask Sensor Dynamic Aperture Mask Static Mask Sensor SIGGRAPH 2007 SIGGRAPH 2007 This Paper Digital Refocusing Light Field Capture
69 Coded Aperture Static Aperture Mask Sensor Optical Heterodyning Static Mask Sensor Reinterpretable Camera Dynamic Aperture Mask Static Mask Sensor SIGGRAPH 2007 SIGGRAPH 2007 Eurographics 2010 Digital Refocusing Light Field Capture Post-Capture Resolution Control
70 Implementation Camera Motor Wheel Shutter Aperture Mask on Wheel Near-Sensor Mask
71 Captured Photo
72 Static Object (in-focus)
73 Static Objects (Out-of-focus)
74 Moving Object (in depth)
75 Rotating Object (in focus)
76 Reconstructed Sub-Aperture Views (3 by 3 Light Field)
77 For Static Objects Angle Angle
78 For Moving Toy in Middle Angle Time
79 For Rotating Toy on Right Time Time
80 High Resolution Image Refocused on Static Toy
81 Digital Refocusing on Static Objects
82 Digital Refocusing on Static Objects
83 Digital Refocusing on Static Objects
84 Digital Refocusing on Static Objects
85 Digital Refocusing on Static Objects
86 Digital Refocusing on Static Objects
87 Digital Refocusing on Toy Moving in Depth
88 Digital Refocusing on Toy Moving in Depth
89 Digital Refocusing on Toy Moving in Depth
90 Digital Refocusing on Toy Moving in Depth
91 Digital Refocusing on Toy Moving in Depth
92 Digital Refocusing on Toy Moving in Depth
93 Video Video for frames Rotating of Toy in-focus
94 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras Camera Coding/Modulation Dimension Flutter Shutter Time (Exposure) Coded Aperture Space Light Field Camera Space and Angle Reinterpretable Camera Space, Time, Angle Flexible Voxels Space and Time
95 Mitsubishi Electric Research Labs (MERL) Computational Cameras Flexible Voxels Similar idea as Reintepretable Camera But for videos Traditional Video Camera Spatial/Temporal Resolution is fixed Scene Independent Flexible Voxels Motion Aware Video Camera Scene dependent variable resolution
96 Sampling of the Space-Time Volume Conventional Sampling Scheme: Sensor Plane Frame 1 Frame 2 Frame N Camera Integration Time Time Our Sampling Scheme: Frame 1 Frame 2 Frame N Camera
97 Co-located Projector-Camera Setup Scene Camera Integration Time Projector Pattern Beam Splitter Image Plane Image Plane Projector Pixel 1 Pixel 2 Pixel K Camera Time 100
98 Multiple Balls Bouncing and Colliding (15 FPS) Close-up Large Motion Blur 101
99 Motion-aware Video Increasing Temporal Resolution + + Captured Frame Different Spatio-temporal Interpretations Motion Analysis Optical Motion-Aware Flow Magnitudes Video
100 Multiple Balls Bouncing Input Sequence Motion-Aware Video 104
101 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras Camera Coding/Modulation Dimension Flutter Shutter Time (Exposure) Coded Aperture Space Light Field Camera Space and Angle Reinterpretable Camera Space, Time and Angle Flexible Voxels Space and Time Common Implementation using fast programmable LCD s
102 Light Fields Camera arrays / Hand-held light field cameras Represented by a set of perspective cameras Typically capture narrow field of view (FOV) light field Narrow FOV [Wilburn et al. 05] [Ng et al. 05] y v u x [Georgiev et al. 06] [Veeraraghavan et al. 07] Set of Perspective Cameras
103 Wide FOV Light Field? Normal Wide FOV Images
104 Wide FOV Light Field! Spherical Mirror Array Refractive Sphere Array
105
106 Wide FOV Refocusing (150 x150 )
107 Focus Back
108 Focus Ball
109 Focus Person
110 All-in-Focus
111 Depth Map
112
113 Wide FOV Refocusing (90 x80 )
114 Focus Back
115 Focus Tree
116 Focus Board
117 All-in-Focus
118 Depth Map
119 Refocusing in Traditional Light Field Object A Refocusing Geometry Projection to Refocusing Geometry Object B Real Cameras Refocus Viewpoint Efficient operation using projective texture mapping on GPU
120 Axial-Cone Modeling of Spherical Mirror Array Real Camera Virtual Cameras Spherical Mirrors
121 Axial-Cone Modeling of Rotationally Symmetric Mirror Real Camera d Captured Photo A cone of rays in the real camera (Angle ) Virtual Camera Rotationally Symmetric Mirror A cone of rays in a virtual camera (Distance d, Angle )
122 Axial-Cone Modeling of Spherical Mirror Array Real Camera Virtual Cameras Spherical Mirrors
123 Axial-Cone Modeling of Refractive Sphere Array Real Camera Refractive Spheres Virtual Cameras
124 Captured Photo Each Sphere Image Axial-Cone Modeling Projection to Refocusing Geometry One Light Field View
125 Light Field Views (100 x100 )
126 Light Field Views (100 x100 )
127 Refocusing Result (100 x100 )
128 Rendering using a Single Perspective Camera Perspective Distortion FOV: 100 x100 FOV: 150 x150
129 Refocusing Result: Cube Map (150 x150 )
130 Refocusing Result: Mercator Projection (150 x150 )
131 Dense Depth Estimation Plane sweeping for dense depth estimation d 3 d 2 d 1 Refocus Viewpoint Check color consistency across light field views at each depth layer
132 Axial-Cones Taguchi, Agrawal, Veeraraghavan, Ramalingam, & Raskar MERL / MIT Media Lab Dense Depth Estimation Plane sweeping for dense depth estimation d 3 d 2 d 1 Refocus Viewpoint Depth Map MITSUBISHI ELECTRIC RESEARCH LABORATORIES
133 Axial-Cones Taguchi, Agrawal, Veeraraghavan, Ramalingam, & Raskar MERL / MIT Media Lab Dense Depth Estimation Plane sweeping for dense depth estimation d 3 d 2 d 1 Refocus Viewpoint All-in-Focus Rendering MITSUBISHI ELECTRIC RESEARCH LABORATORIES
134 Axial-Cones Taguchi, Agrawal, Veeraraghavan, Ramalingam, & Raskar MERL / MIT Media Lab Prototypes Advantages Spherical Mirror Array Single-shot Flexible camera placement Low cost Portable Refractive Sphere Array MITSUBISHI ELECTRIC RESEARCH LABORATORIES
135 Array of 1 Refractive Spheres
136 Refocusing Perspective Projection (90 x80 )
137 All-in-Focus Perspective Projection (90 x80 )
138 Depth Map Perspective Projection (90 x80 )
139 Mitsubishi Electric Research Labs (MERL) Computational Cameras Light Field Mode? Flutter Shutter mode? Reinterpretable Mode?
140 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras Sensing Per-Pixel Control Wide Angle Light Fields Modulation in other dimensions: wavelength Slicing and Sampling of Plentoptic function Reconstruction algorithms Image/video based priors, compressive sensing Statistical properties of plenoptic function
141 Mitsubishi Electric Research Labs (MERL) Computational Cameras Acknowledgements Ramesh Raskar, MIT Jack Tumblin, Northerwestern Univ Ashok Veeraraghavan, Rice Univ. Mohit Gupta, Columbia Univ Ankit Mohan, Flutter Srinivasa Narasimhan, CMU Cyrus Wilson Yuichi Taguchi, MERL Srikumar Ramalingam, MERL MERL, Jay Thornton, Joseph Katz, John Barnwell MELCO, Japan
142 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras Flutter Shutter Camera Coded Aperture Mask based light field camera Reinterpretable Camera Wide Angle light field camera
143 Mitsubishi Electric Research Labs (MERL) Computational Cameras Computational Cameras Flutter Shutter Camera Coded Aperture Mask based light field camera Reinterpretable Camera Wide Angle light field camera
Coding and Modulation in Cameras
Mitsubishi Electric Research Laboratories Raskar 2007 Coding and Modulation in Cameras Ramesh Raskar with Ashok Veeraraghavan, Amit Agrawal, Jack Tumblin, Ankit Mohan Mitsubishi Electric Research Labs
More informationemerging-technology Home > News > emerging-technology > Mitsubishi Electric Develops Deblurring Flutter Shutter Camera
Page 1 of 6 Powershot SD600 Digital $201.00 Powershot S3 IS Digital $319.95 EOS Rebel XTi Black SLR $720.00 PowerShot SD700 IS $259.95 Ads by Google Camera Lens Fujifilm Camera Digital Movie Camera Camera
More informationRadiance Photography. Todor Georgiev Adobe Systems. Andrew Lumsdaine Indiana University
Radiance Photography Todor Georgiev Adobe Systems Andrew Lumsdaine Indiana University Course Goals Overview of radiance (aka lightfield) photography Mathematical treatment of theory and computation Hands
More informationFocal stacks and lightfields
Focal stacks and lightfields http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 11 Course announcements Homework 3 is out. - Due October 12 th.
More informationTutorial: Compressive Sensing of Video
Tutorial: Compressive Sensing of Video Venue: CVPR 2012, Providence, Rhode Island, USA June 16, 2012 Richard Baraniuk Mohit Gupta Aswin Sankaranarayanan Ashok Veeraraghavan Tutorial: Compressive Sensing
More informationComputational Photography
Computational Photography Matthias Zwicker University of Bern Fall 2010 Today Light fields Introduction Light fields Signal processing analysis Light field cameras Application Introduction Pinhole camera
More informationModeling Light. Michal Havlik : Computational Photography Alexei Efros, CMU, Fall 2007
Modeling Light Michal Havlik 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 The Plenoptic Function Figure by Leonard McMillan Q: What is the set of all things that we can ever see? A: The
More informationComputational Photography
Computational Photography Photography and Imaging Michael S. Brown Brown - 1 Part 1 Overview Photography Preliminaries Traditional Film Imaging (Camera) Part 2 General Imaging 5D Plenoptic Function (McMillan)
More informationImage-Based Rendering
Image-Based Rendering COS 526, Fall 2016 Thomas Funkhouser Acknowledgments: Dan Aliaga, Marc Levoy, Szymon Rusinkiewicz What is Image-Based Rendering? Definition 1: the use of photographic imagery to overcome
More informationComputational Photography: Real Time Plenoptic Rendering
Computational Photography: Real Time Plenoptic Rendering Andrew Lumsdaine, Georgi Chunev Indiana University Todor Georgiev Adobe Systems Who was at the Keynote Yesterday? 2 Overview Plenoptic cameras Rendering
More informationNext Generation CAT System
Next Generation CAT System Jaewon Kim MIT Media Lab Advisor: Ramesh Raskar Associate Professor of Media Arts & Sciences MIT Media Lab Reader: V. Michael Bove Principal Research Scientist of Media Arts
More informationA Theory of Multi-Layer Flat Refractive Geometry
A Theory of Multi-Layer Flat Refractive Geometry Axis Amit Agrawal Srikumar Ramalingam Yuichi Taguchi Visesh Chari Mitsubishi Electric Research Labs (MERL) INRIA Imaging with Refractions Source: Shortis
More informationThe Focused Plenoptic Camera. Lightfield photographers, focus your cameras! Karl Marx
The Focused Plenoptic Camera Lightfield photographers, focus your cameras! Karl Marx Plenoptic Camera, Adelson 1992 Main lens focused on microlenses Plenoptic Camera, Adelson 1992 Microlenses focused on
More informationFlexible Voxels for Motion-Aware Videography
Flexible Voxels for Motion-Aware Videography Mohit Gupta 1, Amit Agrawal 2, Ashok Veeraraghavan 2, and Srinivasa G. Narasimhan 1 1 Robotics Institute, Carnegie Mellon University, Pittsburgh, USA 2 Mitsubishi
More informationPlenoptic camera and its Applications
Aum Sri Sairam Plenoptic camera and its Applications Agenda: 1. Introduction 2. Single lens stereo design 3. Plenoptic camera design 4. Depth estimation 5. Synthetic refocusing 6. Fourier slicing 7. References
More informationComputational Imaging for Self-Driving Vehicles
CVPR 2018 Computational Imaging for Self-Driving Vehicles Jan Kautz--------Ramesh Raskar--------Achuta Kadambi--------Guy Satat Computational Imaging for Self-Driving Vehicles Jan Kautz--------Ramesh Raskar--------Achuta
More informationProjective Geometry and Camera Models
/2/ Projective Geometry and Camera Models Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Note about HW Out before next Tues Prob: covered today, Tues Prob2: covered next Thurs Prob3:
More informationComputational Methods for Radiance. Render the full variety offered by the direct observation of objects. (Computationally).
Computational Methods for Radiance Render the full variety offered by the direct observation of objects. (Computationally). Methods for Plenoptic 1.0 Computing with Radiance Goal: Render the full variety
More informationImage-based Lighting
Image-based Lighting 10/17/15 T2 Computational Photography Derek Hoiem, University of Illinois Many slides from Debevec, some from Efros Next week Derek away for ICCV (Venice) Zhizhong and Aditya will
More information3D Shape and Indirect Appearance By Structured Light Transport
3D Shape and Indirect Appearance By Structured Light Transport CVPR 2014 - Best paper honorable mention Matthew O Toole, John Mather, Kiriakos N. Kutulakos Department of Computer Science University of
More informationDEPTH, STEREO AND FOCUS WITH LIGHT-FIELD CAMERAS
DEPTH, STEREO AND FOCUS WITH LIGHT-FIELD CAMERAS CINEC 2014 Frederik Zilly, Head of Group Computational Imaging & Algorithms Moving Picture Technologies Department, Fraunhofer IIS Fraunhofer, Frederik
More informationVIDEO FOR VIRTUAL REALITY LIGHT FIELD BASICS JAMES TOMPKIN
VIDEO FOR VIRTUAL REALITY LIGHT FIELD BASICS JAMES TOMPKIN WHAT IS A LIGHT FIELD? Light field seems to have turned into a catch-all term for many advanced camera/display technologies. WHAT IS A LIGHT FIELD?
More informationAS the most important medium for people to perceive
JOURNAL OF L A T E X CLASS FILES, VOL. XX, NO. X, OCTOBER 2017 1 Light Field Image Processing: An Overview Gaochang Wu, Belen Masia, Adrian Jarabo, Yuchen Zhang, Liangyong Wang, Qionghai Dai, Senior Member,
More informationhttp://www.diva-portal.org This is the published version of a paper presented at 2018 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON), Stockholm Helsinki Stockholm,
More informationChapter 12-Close-Up and Macro Photography
Chapter 12-Close-Up and Macro Photography Close-up images all taken with Hipstamatic on the iphone Close-up and Macro Examples All taken with close-up filter or reverse mount lens Macro Close-up Photography
More informationA projector-camera setup for geometry-invariant frequency demultiplexing
A projector-camera setup for geometry-invariant frequency demultiplexing The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation
More information(54) DIGITAL REFOCUSING FOR WIDE-ANGLE (52) US. Cl /36; 348/E IMAGES USING AXIAL-CONE CAMERAS
US 20110316968A1 (19) United States (12) Patent Application Publication (10) Pub. N0.: US 2011/0316968 A1 Taguchi et a]. (43) Pub. Date: Dec. 29, 2011 (54) DIGITAL REFOCUSING FOR WIDE-ANGLE (52) US. Cl......
More informationCS201 Computer Vision Lect 4 - Image Formation
CS201 Computer Vision Lect 4 - Image Formation John Magee 9 September, 2014 Slides courtesy of Diane H. Theriault Question of the Day: Why is Computer Vision hard? Something to think about from our view
More informationCompensating for Motion During Direct-Global Separation
Compensating for Motion During Direct-Global Separation Supreeth Achar, Stephen T. Nuske, and Srinivasa G. Narasimhan Robotics Institute, Carnegie Mellon University Abstract Separating the direct and global
More informationEfficient Depth-Compensated Interpolation for Full Parallax Displays
ETN-FPI TS3 Plenoptic Sensing Efficient Depth-Compensated Interpolation for Full Parallax Displays Reinhard Koch and Daniel Jung Multimedia Information Processing Full parallax autostereoscopic display
More informationJingyi Yu CISC 849. Department of Computer and Information Science
Digital Photography and Videos Jingyi Yu CISC 849 Light Fields, Lumigraph, and Image-based Rendering Pinhole Camera A camera captures a set of rays A pinhole camera captures a set of rays passing through
More informationHemispherical confocal imaging using turtleback reflector
Hemispherical confocal imaging using turtleback reflector The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published Publisher
More informationAcquiring 4D Light Fields of Self-Luminous Light Sources Using Programmable Filter
Acquiring 4D Light Fields of Self-Luminous Light Sources Using Programmable Filter Motohiro Nakamura 1, Takahiro Okabe 1, and Hendrik P. A. Lensch 2 1 Kyushu Institute of Technology 2 Tübingen University
More informationCS 498 VR. Lecture 20-4/11/18. go.illinois.edu/vrlect20
CS 498 VR Lecture 20-4/11/18 go.illinois.edu/vrlect20 Review from last lecture Texture, Normal mapping Three types of optical distortion? How does texture mipmapping work? Improving Latency and Frame Rates
More informationThe 7d plenoptic function, indexing all light.
Previous Lecture The 7d plenoptic function, indexing all light. Lightfields: a 4d (not 5d!) data structure which captures all outgoing light from a region and permits reconstruction of arbitrary synthetic
More informationLenses. Digital Cameras. Lenses. Lenses. Light focused by the lens. Lenses. Aperture Shutter Speed Holding a Camera Steady Shooting Modes ISO
Lenses Digital Cameras Light focused by the lens Lenses Aperture Shutter Speed Holding a Camera Steady Shooting Modes ISO Lenses Lenses Positive lens Converging lens Focuses image What is difference in
More informationA Theory of Plenoptic Multiplexing
A Theory of Plenoptic Multiplexing Ivo Ihrke ivoihrke@cs.ubc.ca The University of British Columbia ( ) now at Saarland University Gordon Wetzstein wetzste1@cs.ubc.ca The University of British Columbia
More informationTitle: The Future of Photography is Computational Photography. Subtitle: 100 years after invention, Integral Photography becomes feasible
Title: The Future of Photography is Computational Photography Subtitle: 100 years after invention, Integral Photography becomes feasible Adobe customers are creative and demanding. They expect to use our
More informationModeling Light. Slides from Alexei A. Efros and others
Project 3 Results http://www.cs.brown.edu/courses/cs129/results/proj3/jcmace/ http://www.cs.brown.edu/courses/cs129/results/proj3/damoreno/ http://www.cs.brown.edu/courses/cs129/results/proj3/taox/ Stereo
More informationSingle-view 3D Reconstruction
Single-view 3D Reconstruction 10/12/17 Computational Photography Derek Hoiem, University of Illinois Some slides from Alyosha Efros, Steve Seitz Notes about Project 4 (Image-based Lighting) You can work
More informationCS 563 Advanced Topics in Computer Graphics Camera Models. by Kevin Kardian
CS 563 Advanced Topics in Computer Graphics Camera Models by Kevin Kardian Introduction Pinhole camera is insufficient Everything in perfect focus Less realistic Different camera models are possible Create
More informationECE-161C Cameras. Nuno Vasconcelos ECE Department, UCSD
ECE-161C Cameras Nuno Vasconcelos ECE Department, UCSD Image formation all image understanding starts with understanding of image formation: projection of a scene from 3D world into image on 2D plane 2
More informationDD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication
DD2423 Image Analysis and Computer Vision IMAGE FORMATION Mårten Björkman Computational Vision and Active Perception School of Computer Science and Communication November 8, 2013 1 Image formation Goal:
More informationOverview of Active Vision Techniques
SIGGRAPH 99 Course on 3D Photography Overview of Active Vision Techniques Brian Curless University of Washington Overview Introduction Active vision techniques Imaging radar Triangulation Moire Active
More informationIntroduction to 3D Machine Vision
Introduction to 3D Machine Vision 1 Many methods for 3D machine vision Use Triangulation (Geometry) to Determine the Depth of an Object By Different Methods: Single Line Laser Scan Stereo Triangulation
More informationCompressive Sensing of High-Dimensional Visual Signals. Aswin C Sankaranarayanan Rice University
Compressive Sensing of High-Dimensional Visual Signals Aswin C Sankaranarayanan Rice University Interaction of light with objects Reflection Fog Volumetric scattering Human skin Sub-surface scattering
More information3D object recognition used by team robotto
3D object recognition used by team robotto Workshop Juliane Hoebel February 1, 2016 Faculty of Computer Science, Otto-von-Guericke University Magdeburg Content 1. Introduction 2. Depth sensor 3. 3D object
More informationModeling Light. Michal Havlik
Modeling Light Michal Havlik 15-463: Computational Photography Alexei Efros, CMU, Spring 2010 What is light? Electromagnetic radiation (EMR) moving along rays in space R( ) is EMR, measured in units of
More informationPlenoptic Cameras. Bastian Goldlücke, Oliver Klehm, Sven Wanner, and Elmar Eisemann. 5.1 Introduction
Plenoptic Cameras Bastian Goldlücke, Oliver Klehm, Sven Wanner, and Elmar Eisemann 5.1 Introduction The light field, as defined by Gershun in 1936 [Gershun 36] describes the radiance traveling in every
More informationModeling Light. Michal Havlik : Computational Photography Alexei Efros, CMU, Fall 2011
Modeling Light Michal Havlik 15-463: Computational Photography Alexei Efros, CMU, Fall 2011 What is light? Electromagnetic radiation (EMR) moving along rays in space R(λ) is EMR, measured in units of power
More informationI have a meeting with Peter Lee and Bob Cosgrove on Wednesday to discuss the future of the cluster. Computer Graphics
Announcements Assignment 4 will be out later today Problem Set 3 is due today or tomorrow by 9am in my mail box (4 th floor NSH) How are the machines working out? I have a meeting with Peter Lee and Bob
More informationA Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India
A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India Keshav Mahavidyalaya, University of Delhi, Delhi, India Abstract
More informationProjective Geometry and Camera Models
Projective Geometry and Camera Models Computer Vision CS 43 Brown James Hays Slides from Derek Hoiem, Alexei Efros, Steve Seitz, and David Forsyth Administrative Stuff My Office hours, CIT 375 Monday and
More informationReal-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images
Real-time Generation and Presentation of View-dependent Binocular Stereo Images Using a Sequence of Omnidirectional Images Abstract This paper presents a new method to generate and present arbitrarily
More informationMotion and Tracking. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)
Motion and Tracking Andrea Torsello DAIS Università Ca Foscari via Torino 155, 30172 Mestre (VE) Motion Segmentation Segment the video into multiple coherently moving objects Motion and Perceptual Organization
More informationMultiple View Geometry
Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V
More informationThree-dimensional directional display and pickup
Three-dimensional directional display and pickup Joonku Hahn NCRCAPAS School of Electrical Engineering Seoul National University CONTENTS I. Introduction II. Three-dimensional directional display 2.1 Uniform
More informationScalable multi-gpu cloud raytracing with OpenGL
Scalable multi-gpu cloud raytracing with OpenGL University of Žilina Digital technologies 2014, Žilina, Slovakia Overview Goals Rendering distant details in visualizations Raytracing Multi-GPU programming
More informationMohit Gupta. Assistant Professor, Computer Science Department University of Wisconsin-Madison. Ph.D. in Robotics, Carnegie Mellon University
Mohit Gupta Assistant Professor, Computer Science Department University of Wisconsin-Madison Email: mohitg@cs.wisc.edu http://pages.cs.wisc.edu/ mohitg Research Interests Computer Vision, Computational
More informationHTC ONE (M8) DUO CAMERA WHITE PAPERS
HTC ONE (M8) DUO CAMERA WHITE PAPERS 1 HTC ONE (M8) DUO CAMERA WHITE PAPERS INTRODUCTION In a few short years, the smartphone has become the predominant image-capturing device for millions of people. At
More informationLight. Properties of light. What is light? Today What is light? How do we measure it? How does light propagate? How does light interact with matter?
Light Properties of light Today What is light? How do we measure it? How does light propagate? How does light interact with matter? by Ted Adelson Readings Andrew Glassner, Principles of Digital Image
More informationPinhole Camera Model 10/05/17. Computational Photography Derek Hoiem, University of Illinois
Pinhole Camera Model /5/7 Computational Photography Derek Hoiem, University of Illinois Next classes: Single-view Geometry How tall is this woman? How high is the camera? What is the camera rotation? What
More informationRigid Body Motion and Image Formation. Jana Kosecka, CS 482
Rigid Body Motion and Image Formation Jana Kosecka, CS 482 A free vector is defined by a pair of points : Coordinates of the vector : 1 3D Rotation of Points Euler angles Rotation Matrices in 3D 3 by 3
More informationDepth Estimation with a Plenoptic Camera
Depth Estimation with a Plenoptic Camera Steven P. Carpenter 1 Auburn University, Auburn, AL, 36849 The plenoptic camera is a tool capable of recording significantly more data concerning a particular image
More informationTime-of-flight basics
Contents 1. Introduction... 2 2. Glossary of Terms... 3 3. Recovering phase from cross-correlation... 4 4. Time-of-flight operating principle: the lock-in amplifier... 6 5. The time-of-flight sensor pixel...
More informationOther approaches to obtaining 3D structure
Other approaches to obtaining 3D structure Active stereo with structured light Project structured light patterns onto the object simplifies the correspondence problem Allows us to use only one camera camera
More informationToday. Global illumination. Shading. Interactive applications. Rendering pipeline. Computergrafik. Shading Introduction Local shading models
Computergrafik Matthias Zwicker Universität Bern Herbst 2009 Today Introduction Local shading models Light sources strategies Compute interaction of light with surfaces Requires simulation of physics Global
More informationL1a: Introduction to Light Fields
L1a: Introduction to Light Fields 2018 IEEE SPS Summer School on Light Field Data Representation, Interpretation, and Compression Donald G. Dansereau, May 2018 Schedule 2 Outline Lecture 1a: Introduction
More informationCapturing light. Source: A. Efros
Capturing light Source: A. Efros Review Pinhole projection models What are vanishing points and vanishing lines? What is orthographic projection? How can we approximate orthographic projection? Lenses
More informationConsider a partially transparent object that is illuminated with two lights, one visible from each side of the object. Start with a ray from the eye
Ray Tracing What was the rendering equation? Motivate & list the terms. Relate the rendering equation to forward ray tracing. Why is forward ray tracing not good for image formation? What is the difference
More informationMAS.963 Special Topics: Computational Camera and Photography
MIT OpenCourseWare http://ocw.mit.edu MAS.963 Special Topics: Computational Camera and Photography Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationLight Field Spring
Light Field 2015 Spring Recall: Light is Electromagnetic radiation (EMR) moving along rays in space R(l) is EMR, measured in units of power (watts) l is wavelength Useful things: Light travels in straight
More informationVideo Mosaics for Virtual Environments, R. Szeliski. Review by: Christopher Rasmussen
Video Mosaics for Virtual Environments, R. Szeliski Review by: Christopher Rasmussen September 19, 2002 Announcements Homework due by midnight Next homework will be assigned Tuesday, due following Tuesday.
More informationThe 2 nd part of the photographic triangle
The 2 nd part of the photographic triangle Shutter speed refers to the amount of time your sensor is exposed to light. In film photography shutter speed was the length of time that the film was exposed
More informationImage Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania
Image Formation Antonino Furnari Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania furnari@dmi.unict.it 18/03/2014 Outline Introduction; Geometric Primitives
More informationP2C2: Programmable Pixel Compressive Camera for High Speed Imaging
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com P2C2: Programmable Pixel Compressive Camera for High Speed Imaging Reddy, D.; Veeraraghavan, A.; Chellappa, R. TR2011-039 June 2011 Abstract
More informationAN O(N 2 LOG(N)) PER PLANE FAST DISCRETE FOCAL STACK TRANSFORM
AN O(N 2 LOG(N)) PER PLANE FAST DISCRETE FOCAL STACK TRANSFORM Fernando Pérez Nava +, Jonás Philipp Lüke + Departamento de Estadística, Investigación Operativa y Computación Departamento de Física Fundamental
More informationAll human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them.
All human beings desire to know. [...] sight, more than any other senses, gives us knowledge of things and clarifies many differences among them. - Aristotle University of Texas at Arlington Introduction
More informationFlat-Field Mega-Pixel Lens Series
Flat-Field Mega-Pixel Lens Series Flat-Field Mega-Pixel Lens Flat-Field Mega-Pixel Lens 205.ver.0 E Specifications and Lineup Full Full Full Full 5MP MP MP MP Image Model Imager Size Mount Focal Length
More informationAnnouncements. Written Assignment 2 out (due March 8) Computer Graphics
Announcements Written Assignment 2 out (due March 8) 1 Advanced Ray Tracing (Recursive) Ray Tracing Antialiasing Motion Blur Distribution Ray Tracing Ray Tracing and Radiosity Assumptions Simple shading
More information3D Surface Reconstruction Based On Plenoptic Image. Haiqiao Zhang
3D Surface Reconstruction Based On Plenoptic Image by Haiqiao Zhang A thesis submitted to the Graduate Faculty of Auburn University in partial fulfillment of the requirements for the Degree of Master of
More informationCMSC427 Shading Intro. Credit: slides from Dr. Zwicker
CMSC427 Shading Intro Credit: slides from Dr. Zwicker 2 Today Shading Introduction Radiometry & BRDFs Local shading models Light sources Shading strategies Shading Compute interaction of light with surfaces
More information3D Time-of-Flight Image Sensor Solutions for Mobile Devices
3D Time-of-Flight Image Sensor Solutions for Mobile Devices SEMICON Europa 2015 Imaging Conference Bernd Buxbaum 2015 pmdtechnologies gmbh c o n f i d e n t i a l Content Introduction Motivation for 3D
More informationIntroduction to Shutter Speed in Digital Photography. Read more:
Introduction to Shutter Speed in Digital Photography Read more: http://digital-photography-school.com/shutterspeed#ixzz26mrybgum What is Shutter Speed? shutter speed is the amount of time that the shutter
More informationPerceptual Effects in Real-time Tone Mapping
Perceptual Effects in Real-time Tone Mapping G. Krawczyk K. Myszkowski H.-P. Seidel Max-Planck-Institute für Informatik Saarbrücken, Germany SCCG 2005 High Dynamic Range (HDR) HDR Imaging Display of HDR
More informationMultiview Reconstruction
Multiview Reconstruction Why More Than 2 Views? Baseline Too short low accuracy Too long matching becomes hard Why More Than 2 Views? Ambiguity with 2 views Camera 1 Camera 2 Camera 3 Trinocular Stereo
More informationMinimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc.
Minimizing Noise and Bias in 3D DIC Correlated Solutions, Inc. Overview Overview of Noise and Bias Digital Image Correlation Background/Tracking Function Minimizing Noise Focus Contrast/Lighting Glare
More informationComputer Graphics. Lecture 9 Environment mapping, Mirroring
Computer Graphics Lecture 9 Environment mapping, Mirroring Today Environment Mapping Introduction Cubic mapping Sphere mapping refractive mapping Mirroring Introduction reflection first stencil buffer
More informationDigital Imaging Study Questions Chapter 8 /100 Total Points Homework Grade
Name: Class: Date: Digital Imaging Study Questions Chapter 8 _/100 Total Points Homework Grade True/False Indicate whether the sentence or statement is true or false. 1. You can change the lens on most
More informationIntroduction to Photography
Introduction to Photography The Camera Digital Cameras The Camera (front & top) The Camera (back & bottom) Digital Camera Modes Scene Modes Landscape Photography What makes a good landscape? http://photography.nationalgeographic.com/phot
More informationUsing Your Digital Camera
Using Your Digital Camera Presented by Tallahassee Senior Center Volunteer Instructor: Dr. Mike Francis Email: drmikef@comcast.net http://www.maf1.com/dcc 1 Digital Camera Topics for Everyone Camera Components
More informationVolumetric Scene Reconstruction from Multiple Views
Volumetric Scene Reconstruction from Multiple Views Chuck Dyer University of Wisconsin dyer@cs cs.wisc.edu www.cs cs.wisc.edu/~dyer Image-Based Scene Reconstruction Goal Automatic construction of photo-realistic
More informationINFOGR Computer Graphics. J. Bikker - April-July Lecture 10: Ground Truth. Welcome!
INFOGR Computer Graphics J. Bikker - April-July 2015 - Lecture 10: Ground Truth Welcome! Today s Agenda: Limitations of Whitted-style Ray Tracing Monte Carlo Path Tracing INFOGR Lecture 10 Ground Truth
More informationReferences Photography, B. London and J. Upton Optics in Photography, R. Kingslake The Camera, The Negative, The Print, A. Adams
Page 1 Camera Simulation Eect Cause Field o view Depth o ield Motion blur Exposure Film size, stops and pupils Aperture, ocal length Shutter Film speed, aperture, shutter Reerences Photography, B. London
More informationModeling Light. Michal Havlik
Modeling Light Michal Havlik 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 What is light? Electromagnetic radiation (EMR) moving along rays in space R(λ) is EMR, measured in units of power
More informationGenesis : The first complete 35mm Digital Cinematography System
Genesis : The first complete 35mm Digital Cinematography System GENESIS & Super35 16:9 film image size compared to 2/3" CCD image size 3 perf. 16:9 35 mm film camera apertur e 13.5mm (13.87mm) 2/3" CC
More informationDEPTH AND ANGULAR RESOLUTION IN PLENOPTIC CAMERAS. M. Damghanian, R. Olsson, M. Sjöström
This material is published in the open archive of Mid Sweden University DIVA http://miun.diva-portal.org to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein
More informationWe ll go over a few simple tips for digital photographers.
Jim West We ll go over a few simple tips for digital photographers. We ll spend a fair amount of time learning the basics of photography and how to use your camera beyond the basic full automatic mode.
More informationRadiometry and reflectance
Radiometry and reflectance http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 16 Course announcements Homework 4 is still ongoing - Any questions?
More informationINTRODUCTION TO MEDICAL IMAGING- 3D LOCALIZATION LAB MANUAL 1. Modifications for P551 Fall 2013 Medical Physics Laboratory
INTRODUCTION TO MEDICAL IMAGING- 3D LOCALIZATION LAB MANUAL 1 Modifications for P551 Fall 2013 Medical Physics Laboratory Introduction Following the introductory lab 0, this lab exercise the student through
More information