Comment on Numerical shape from shading and occluding boundaries

Similar documents
DIRECT SHAPE FROM ISOPHOTES

Shape from shading. Surface brightness and Surface Orientation --> Reflectance map READING: Nalwa Chapter 5. BKP Horn, Chapter 10.

General Principles of 3D Image Analysis

Passive 3D Photography

A Survey of Light Source Detection Methods

A Shape from Shading Approach for the Reconstruction of Polyhedral Objects using Genetic Algorithm

Reconstruction of Discrete Surfaces from Shading Images by Propagation of Geometric Features

Photometric Stereo. Lighting and Photometric Stereo. Computer Vision I. Last lecture in a nutshell BRDF. CSE252A Lecture 7

MULTIRESOLUTION SHAPE FROM SHADING*

Investigation of Directional Filter on Kube-Pentland s 3D Surface Reflectance Model using Photometric Stereo

Last update: May 4, Vision. CMSC 421: Chapter 24. CMSC 421: Chapter 24 1

COMP 558 lecture 16 Nov. 8, 2010

SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD

CAN THE SUN S DIRECTION BE ESTIMATED PRIOR TO THE DETERMINATION OF SHAPE?

Silhouette-based Multiple-View Camera Calibration

Motion Estimation. There are three main types (or applications) of motion estimation:

3D Motion Analysis Based on 2D Point Displacements

Surface reconstruction from photometric stereo images

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Announcement. Lighting and Photometric Stereo. Computer Vision I. Surface Reflectance Models. Lambertian (Diffuse) Surface.

CS 4495 Computer Vision Motion and Optic Flow

Integrating Shape from Shading and Shape from Stereo for Variable Reflectance Surface Reconstruction from SEM Images

Shadow Graphs and Surface Reconstruction

Comparison Between The Optical Flow Computational Techniques

Image Processing 1 (IP1) Bildverarbeitung 1

Photometric Stereo, Shape from Shading SfS Chapter Szelisky

Image segmentation through reasoning about optics

Motion Analysis. Motion analysis. Now we will talk about. Differential Motion Analysis. Motion analysis. Difference Pictures

Analysis of photometric factors based on photometric linearization

DD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication

EE795: Computer Vision and Intelligent Systems

Feature Tracking and Optical Flow

Photometric Stereo.

Occluded Facial Expression Tracking

Computational Perception /785

Variational Multiframe Stereo in the Presence of Specular Reflections

Passive 3D Photography

Shading and Recognition OR The first Mrs Rochester. D.A. Forsyth, UIUC

Notes 9: Optical Flow

Chapter 7. Conclusions and Future Work

Optimal Grouping of Line Segments into Convex Sets 1

Shape from Intensity Gradient

Towards the completion of assignment 1

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Polarization-based Transparent Surface Modeling from Two Views

A Statistical Consistency Check for the Space Carving Algorithm.

Peripheral drift illusion

June Wesley, Cambridge Univ. Press,

Stereo: Disparity and Matching

From a real chair to a negative chair

3-D Stereo Using Photometric Ratios

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Lambertian model of reflectance I: shape from shading and photometric stereo. Ronen Basri Weizmann Institute of Science

ASSOCIATIVE LEARNING OF STANDARD REGULARIZING OPERATORS IN EARLY VISION

Homework 4 Computer Vision CS 4731, Fall 2011 Due Date: Nov. 15, 2011 Total Points: 40

What is Computer Vision? Introduction. We all make mistakes. Why is this hard? What was happening. What do you see? Intro Computer Vision

Three-Dimensional Computer Vision

Particle Tracking. For Bulk Material Handling Systems Using DEM Models. By: Jordan Pease

Announcement. Photometric Stereo. Computer Vision I. Shading reveals 3-D surface geometry. Shape-from-X. CSE252A Lecture 8

Calculating the Surface Topography of Integrated Circuit Wafers from SEM Images

Other Reconstruction Techniques

High Quality Shape from a Single RGB-D Image under Uncalibrated Natural Illumination

Computer Vision Lecture 17

PART-LEVEL OBJECT RECOGNITION

CS 565 Computer Vision. Nazar Khan PUCIT Lectures 15 and 16: Optic Flow

Computer Vision Lecture 17

Photometric Stereo. Photometric Stereo. Shading reveals 3-D surface geometry BRDF. HW3 is assigned. An example of photometric stereo

Estimation of Illuminant Direction, Albedo, and Shape from Shading*

Capturing, Modeling, Rendering 3D Structures

Understanding Variability

Limits on the Perception of Local Shape from Shading

Skeleton Cube for Lighting Environment Estimation

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Scale Space Surface Recovery Using Binocular Shading and Stereo Information

Motion and Tracking. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

VC 11/12 T11 Optical Flow

Overview of Active Vision Techniques

CEng Computational Vision

High Accuracy Depth Measurement using Multi-view Stereo

Feature Tracking and Optical Flow

Lecture 10: Multi view geometry

Finally: Motion and tracking. Motion 4/20/2011. CS 376 Lecture 24 Motion 1. Video. Uses of motion. Motion parallax. Motion field

Hand-Eye Calibration from Image Derivatives

Textureless Layers CMU-RI-TR Qifa Ke, Simon Baker, and Takeo Kanade

Height from Gradient with Surface Curvature and Area Constraints

Multi-stable Perception. Necker Cube

Computer Vision. I-Chen Lin, Assistant Professor Dept. of CS, National Chiao Tung University

2 The MiníMax Principle First consider a simple problem. This problem will address the tradeos involved in a two-objective optimiation problem, where

CS6670: Computer Vision

Capturing light. Source: A. Efros

Massachusetts Institute of Technology. Department of Computer Science and Electrical Engineering /6.866 Machine Vision Quiz I

C. Lawrence Zitnick, Jon A. Webb

Structure from Motion. Lecture-15

Announcements. Hough Transform [ Patented 1962 ] Generalized Hough Transform, line fitting. Assignment 2: Due today Midterm: Thursday, May 5 in class

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Final Exam Study Guide

Depth. Chapter Stereo Imaging

A Simple Vision System

Light source estimation using feature points from specular highlights and cast shadows

Marcel Worring Intelligent Sensory Information Systems

Transcription:

Artificial Intelligence 59 (1993) 89-94 Elsevier 89 ARTINT 1001 Comment on Numerical shape from shading and occluding boundaries K. Ikeuchi School of Compurer Science. Carnegie Mellon dniversity. Pirrsburgh. PA 15213. USA A bsiract Ikeuchi. K.. Comment on Numerical shape from shading and occluding boundaries, Artificial Intelligence 59 (1993) 89-94. The paper Numerical shape from shading and occluding boundaries [IO] proposed a method to determine the 3D shape of an object from a single brightness image using the smoothness constraint on surface orientations, and the orientations of the surface given along an occluding boundary. This paper contributed both a specific algorithm for solving a specific problem, and a general approach that could be applied in other low-level vision problems. 1. Research background It has long been known that the human visual system is capable of determining the 3D structure (relative depth or surface orientations) of a scene, given only 2D information. For example, a black-and-white photograph of an object is only 2D, yet can be used by a human to determine the 3D shape of the object. The human visual system uses a variety of cues to determine 3D shape. One such cue is shading information: the way in which brightness varies along the surface of an object is used by the human visual system to determine the underlying 3D shape of the object; the process is known as shape-from-shading. Correspondence 10: K. Ikeuchi, School of Computer Science, Carnegie Mellon University. Pittsburgh, PA 15213. USA. 0004-3702/93/% 06.00 @ 1993 - Elsevier Science Publishers B.V.,411 rights reserved

90 K. Ikeuchi In the late 1970s and early 1980s, at the Artificial Intelligence Laboratory of the Massachusetts Institute of Technology, several projects were conducted to investigate computational algorithms to emulate on computers these modules of the human visual system that determine (relative) depth from images. Such projects investigated shape-from-shading (Horn, Woodham, Ikeuchi, Pentland), shape-from-texture (Richard, Witkin, Stevens ), optical flow (Horn, Marr, Schunk, Hildreth), and binocular stereo (Marr, Poggio, Grimson). Shape-from-shading research, originally suggested as a thesis topic by Minsky, was initiated by Horn as his Ph.D. thesis in 1970 [ 71. Horn formalized the shape-from-shading problem using the nonlinear partial differential image irradiance equation, where R(p,q) is a known reflectance function of an object surface, represented using the standard (p,q) gradient space coordinate system to define unknown orientations and assuming orthographic projection down the z axis onto the (x,y) plane for image formation, and E(x,y) is a given input brightness distribution at each image point, (x, y). Horn solved the equation using the characteristic strip expansion method. The method draws characteristic strips-special strips which satisfy a certain condition-in a shaded image and determines orientations iteratively along the characteristic strips from a point whose orientation is known. As the starting point, Horn used the point which provides the maximum brightness, R(p, q) = 1. (In particular, on a Lambertian surface, the point of the maximum brightness is oriented toward the light source.) The method demonstrated was able to determine the shape of 3D objects such as a human nose from its black/white shaded image. Horn s original method, however, has two main defects. Direct integration of the characteristic equations to find the characteristic strips suffers from noise sensitivity in practical implementations. As the solution proceeds along constructed strips from the initial points, these strips, as well as the computed orientations, can deviate from the actual strip positions and orientations as a result of quantization errors and other noise influences. 0 The method starts characteristic strips from a point with a known orientation. The method employs the point of maximum brightness for the starting point. However, at such points the partial derivatives of R vanish, so the strips cannot be uniquely constructed. To overcome this problem, the original method uses the points on a circumscribing circle to approximate the starting points.

Comment on Numerical shape from shading and occluding boundaries 91 2. Technical contents In 1979 we (Ikeuchi and Horn) solved these problems by introducing the smoothness constraint and by utilizing the information given along occluding boundaries: 0 Noise sensitivity: Horn s original method can only determine surface orientations along the characteristic strips. Thus, a single noisy point on a strip could throw off all of the following computations. Our method introduced the smoothness constraint, which required neighboring points to have similar surface orientations: P: + P; + 4: + 4; = 0. By using this constraint, our method propagated, combined, and averaged orientations over neighboring grid points to determine the orientation at the central point. By averaging the information in a neighborhood, sensitivity to noise was greatly reduced. 0 Singular point: Horn s original method can only use the orientation at a singular point. Our method uses more widely available orientations along an occluding boundary. (Horn s method cannot use these orientations due to its parameterization.) At the points along occluding boundaries of objects, orientations are perpendicular to the lines of sight. From the shape of the silhouette, which is the projection of an occluding boundary onto the image plane, we can determine the tangential direction of the occluding boundary parallel to the image plane, and thus, the orientation there. Our method uses these orientations as the initial conditions for the algorithm. The image irradiance equation, E(x,y) - R(p,q) = 0, only provides one constraint at each image point. The surface orientation of the point has two degrees of freedom (p,q). We used the smoothness constraint, p: + p; + qi + q: = 0, as an additional constraint. (Horn s method uses a weak smoothness constraint given by the continuity of the characteristics strips. ) We formalized the shape-from-shading problem as a minimization problem of the difference between the observed brightness and the expected brightness given through the image irradiance equation from the expected surface orientation, while maintaining the smoothness constraint. c c (2) The original working memo was published November 1979, the revised AI memo published February 1980, and the revised journal article was published in August 1981.

92 K. Ikeuchi The minimization is performed over the region, A, enclosed by an occluding boundary, da; the aim is to obtain a distribution of (p, 4)s which minimize the equations over the region, A. Using the calculus of variations, we formulated an iterative scheme to obtain the (p,q) which minimizes the equation. Since the iterative scheme requires a boundary condition, we used the orientations along the occluding boundary, aa. The method was demonstrated to determine orientations over the region, A, from the shading information over the region, A, and the orientation along the boundary, da. 3. Contributions One of the main intellectual contributions of the paper came from reporting the effectiveness of the smoothness constraint on low-level vision modules. The smoothness constraint forces neighboring points to have similar surface orientations. Later this constraint was generalized into the regufurization term by Poggio. The paper predicted that the smoothness constraint would be effective with other low-level vision modules. (See the discussion in [IO, p. 1811.) The paper also demonstrated the effectiveness of the variational approach for low-level vision problems. Using this method, various low-level vision problems can be solved in a uniform manner. 4. Open issues The following are research issues that were left out of the original paper but are still open. Some of the recent representative research toward the issues are also discussed: The smoothness constraint Though robust, the method based on the smoothness constraint does deviate from the real solution. That is, the algorithm converges to shapes that are slightly different from the real shapes. Another problem of the smoothness constraint is that it does not guarantee that the solution obtained is integrable because our method determines p and q independently. Several researchers (Horn and Brooks [ 5,6 J and Frankot and Chelappa [ 41 ) pointed out these two effects and proposed new algorithms by introducing integrability constraints. However, I (Ikeuchi) believe that the smoothness constraint is still interesting enough as a mechanism to emulate modules of the human visual system.

Comment on "Numerical shape from shading and occluding boundaries'' 93 (1) The smoothness constraint can be applied to a wide variety of lowlevel vision modules regardless of parameterizations. Some of the proposed constraints are only valid for a certain parameterization. The notion of smoothness should be independent of parameterizations. It is quite likely that the human visual system imposes such smoothness constraints (nearby points have similar characteristics) to its default interpretation of a scene. (2) The method based on the smoothness constraint has the preferred default shape. In the absence of any shading information, the method applied to the boundary condition yields the default shape. When shading information is present, this shape is modified so as to satisfy the image irradiance equation. It seems to me that this mechanism is quite reasonable as the behavior of the human visual system. Namely, the mechanism has an underlying default shape and outputs such a default shape when no input information is available. When some input is available, it modifies the output accordingly. From these two points, I conjecture that some sort of smoothness constraint mechanism exists in the human visual system. Research on the similarities and differences between human and machine shape-from-shading algorithms is an interesting topic. Such comparisons are reported by Mingolla and Todd [11 I. Occluding boundary The method requires surface orientations along a closed occluding boundary. We can observe a whole occluding boundary only when a viewer and a light source are located at the same position. Under general conditions, an illuminated region is partly enclosed by an occluding boundary and partly by a self-shadowed boundary. Along a self-shadowed boundary, no orientation information is available. Thus, under general conditions, our method cannot be applied to the shape-from-shading problem. It is necessary to remedy this point (see Pentland [ 141). Binocular stereo provides depth information along a boundary, aa. In order to obtain depth information over the region A, it is necessary to interpolate depth information from that along the boundary da. Shapefrom-shading provides (relative) depth information over the region A, from those along the boundary da. Both modules are complementary in nature, and thus combining the methods is an interesting research goal (Ikeuchi [8,9], Blake et al. [ 11). Known reflectance map Our shape-from-shading algorithm requires a known exact reflectance map, that is, known reflectance characteristics, and a known light source direction.

94 K. Ikeuchi It is rather unusual to know the exact reflectance characteristics of an object and the exact direction of the light source. A method for determining the shape using a rough reflectance map is necessary (Brooks and Horn 121. Nayar et al. [ 121 ). Interrejlection Finally, the method assumes that one light source directly illuminates a surface. However, it often occurs that several light sources illuminate a surface simultaneously, and interreflections occur among nearby surfaces. Research on shape-from-shading under complicated illumination conditions should be explored (Forsyth [ 31, Nayar et al. [ 131 ). References [ 1 ] A. Blake, A. Zisserman and G. Knowles, Surface descriptions from stereo and shading, Image Vis. Comput. 3 (4) (1985) 183-191. [2] M.J. Brooks and BKP. Horn, Shape and source from shading, in: Proceedings IJCAI-85, Los Angeles, CA (1985) 932-936. [ 3 1 D. Fonyth and A. Zisserman, Reflections on shading, IEEE Trans. Pattern Anal. Mach. fntefl, 13 (7) (1991) 671-679. [4] R.T. Frankot and R. Chellappa, A method for enforcing integrability in shape from shading algorithm, IEEE Trans. Pattern Anal. Mach. Intell. 10 (4) ( 1988 439-45 1. (51 B.K.P. Horn, Height and gradient from shading, Int. J. Comput. Vis. 5 (I) (1990) 37-74. [ 6 J B.K.P. Horn and M.J. Brooks, The variational approach to shape from shading, Compuf. Vis. Graph. Image Process. 33 (2) (1986) 174-208. [7] B.K.P. Horn, Shape from shading: a method for obtaining the shape of a smooth opaque object from one view, Tech. Rept. MAC-TR-79, MIT, Cambridge, MA (1970). [8] K. Ikeuchi, Reconstructing a depth map from intensity map, in: Proceedings International Conference on Pattern Recognition (1984) 736-738. 191 K. Ikeuchi, Determining a depth map using a dual photometric stereo, fnt. J. Rob. Res. 6 (1) (1987) 15-31. [ 101 K. Ikeuchi and B.K.P. Horn, Numerical shape from shading and occluding boundaries, Artif: Intell. 17 (1981) 141-184. [ I 1 ] E. Mingolla and J.T. Todd, Perception of solid shape from shading, Biol. Cybern. 53 (1986) 137-151. [12] S.K. Nayar, K. Ikeuchi and T. Kanade, Determining shape and reflectance of hybrid surfaces by phctomctric sampling, IEEE Trans. Rob. Autom. 6 (4) ( 1990) 418-43 1. [ 131 S.K. Nayar, K. Ikeuchi and T. Kanade, Shape from interreflections, Int. J. Comput. Vis. 6 (3) (1991) 173-195. [ 141 A.P. Pentland, Linear shape from shading, fnt. J. Compuf. Vis. 4 (1990) 153-162.