Efficient illuminant correction in the Local, Linear, Learned (L 3 ) method

Similar documents
EXPERIMENTAL HIGH SPEED CMOS IMAGE SENSOR SYSTEM & APPLICATIONS

Color. Reading: Optional reading: Chapter 6, Forsyth & Ponce. Chapter 4 of Wandell, Foundations of Vision, Sinauer, 1995 has a good treatment of this.

Color. Computational Photography MIT Feb. 14, 2006 Bill Freeman and Fredo Durand

Estimating the wavelength composition of scene illumination from image data is an

WHITE PAPER. Application of Imaging Sphere for BSDF Measurements of Arbitrary Materials

Introduction to color science

Why does a visual system need color? Color. Why does a visual system need color? (an incomplete list ) Lecture outline. Reading: Optional reading:

Digital Image Processing COSC 6380/4393

Computer Vision. The image formation process

Siggraph Course 2017 Path Tracing in Production Part 1 Manuka: Weta Digital's Spectral Renderer

Diagonal versus affine transformations for color correction

An Algorithm to Determine the Chromaticity Under Non-uniform Illuminant

Classifying Body and Surface Reflections using Expectation-Maximization

Physics-based Vision: an Introduction

Miniaturized Camera Systems for Microfactories

Spectral Images and the Retinex Model

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

CS5670: Computer Vision

dependent intensity function - the spectral distribution function (SPD) E( ). The surface reflectance is the proportion of incident light which is ref

CS 563 Advanced Topics in Computer Graphics Film and Image Pipeline (Ch. 8) Physically Based Rendering by Travis Grant.

Lecture 1 Image Formation.

Computer Vision Course Lecture 02. Image Formation Light and Color. Ceyhun Burak Akgül, PhD cba-research.com. Spring 2015 Last updated 04/03/2015

Storage Efficient NL-Means Burst Denoising for Programmable Cameras

Fast Spectral Reflectance Recovery using DLP Projector

CS6670: Computer Vision

Inner Products and Orthogonality in Color Recording Filter Design

Introducing Robotics Vision System to a Manufacturing Robotics Course

FACE DETECTION AND RECOGNITION OF DRAWN CHARACTERS HERMAN CHAU

Combining Strategies for White Balance

CSE 167: Lecture #7: Color and Shading. Jürgen P. Schulze, Ph.D. University of California, San Diego Fall Quarter 2011

EECS 556 Image Processing W 09. Image enhancement. Smoothing and noise removal Sharpening filters

CSE 167: Introduction to Computer Graphics Lecture #6: Colors. Jürgen P. Schulze, Ph.D. University of California, San Diego Fall Quarter 2013

3D object recognition used by team robotto

(Refer Slide Time 00:17) Welcome to the course on Digital Image Processing. (Refer Slide Time 00:22)

Glare Spread Function (GSF) - 12 Source Angle

Announcements. Lighting. Camera s sensor. HW1 has been posted See links on web page for readings on color. Intro Computer Vision.

Calibration of Recorded Digital Counts to Radiance

Spectral characterization of a color scanner by adaptive estimation

IMPORTANT INSTRUCTIONS

Light. Properties of light. What is light? Today What is light? How do we measure it? How does light propagate? How does light interact with matter?

FC Series Traffic Camera. Architect & Engineering Specifications

CINEMA EOS LENSES COLOR REPRODUCTION PERSONALITY OF THE CANON CINEMA EOS LENS: WHITE PAPER

Polynomial Regression Spectra Reconstruction of Arctic Charr s RGB

CS6670: Computer Vision

Mapping of Hierarchical Activation in the Visual Cortex Suman Chakravartula, Denise Jones, Guillaume Leseur CS229 Final Project Report. Autumn 2008.

A Qualitative Analysis of 3D Display Technology

High Information Rate and Efficient Color Barcode Decoding

Scientific imaging of Cultural Heritage: Minimizing Visual Editing and Relighting

Introduction to Remote Sensing Wednesday, September 27, 2017

Lecture #2: Color and Linear Algebra pt.1

And. Modal Analysis. Using. VIC-3D-HS, High Speed 3D Digital Image Correlation System. Indian Institute of Technology New Delhi

An LED based spectrophotometric instrument

Announcements. Camera Calibration. Thin Lens: Image of Point. Limits for pinhole cameras. f O Z

Color. Phillip Otto Runge ( )

On the distribution of colors in natural images

Quantitative Analysis of Metamerism for. Multispectral Image Capture

Modeling and Estimation of FPN Components in CMOS Image Sensors

Geodesic Based Ink Separation for Spectral Printing

Global Illumination. Frank Dellaert Some slides by Jim Rehg, Philip Dutre

Technical Basis for optical experimentation Part #4

Introduction to Digital Image Processing

A Statistical Approach to Culture Colors Distribution in Video Sensors Angela D Angelo, Jean-Luc Dugelay

Color to Binary Vision. The assignment Irfanview: A good utility Two parts: More challenging (Extra Credit) Lighting.

Opponent Color Spaces

EE795: Computer Vision and Intelligent Systems

We ll go over a few simple tips for digital photographers.

Minimizing Noise and Bias in 3D DIC. Correlated Solutions, Inc.

Improvements to Gamut Mapping Colour Constancy Algorithms

Color Vision. Spectral Distributions Various Light Sources

Sensors and Camera Systems for Scientific, Industrial, and Digital Photography Applications II, Morley M. Blouke, John Canosa, Nitin Sampat, Editors,

Statistical image models

Today. Global illumination. Shading. Interactive applications. Rendering pipeline. Computergrafik. Shading Introduction Local shading models

LEVEL 1 ANIMATION ACADEMY2010

Imaging Sphere Measurement of Luminous Intensity, View Angle, and Scatter Distribution Functions

Classifying Hand Shapes on a Multitouch Device

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier

COLOR FIDELITY OF CHROMATIC DISTRIBUTIONS BY TRIAD ILLUMINANT COMPARISON. Marcel P. Lucassen, Theo Gevers, Arjan Gijsenij

CS4670: Computer Vision

Motivation. Intensity Levels

3D Shape and Indirect Appearance By Structured Light Transport

CIE L*a*b* color model

Novel Lossy Compression Algorithms with Stacked Autoencoders

Image Contrast Enhancement in Wavelet Domain

Chemlux Imager. Applications

Vivekananda. Collegee of Engineering & Technology. Question and Answers on 10CS762 /10IS762 UNIT- 5 : IMAGE ENHANCEMENT.

IMAGE ENHANCEMENT in SPATIAL DOMAIN by Intensity Transformations

Mingle Face Detection using Adaptive Thresholding and Hybrid Median Filter

INFINITY-CORRECTED TUBE LENSES

Automatic Fatigue Detection System

Independent Resolution Test of

Rodenstock Products Photo Optics / Digital Imaging

Capturing, Modeling, Rendering 3D Structures

A New Image Based Ligthing Method: Practical Shadow-Based Light Reconstruction

Computational Photography

Estimating the surface normal of artwork using a DLP projector

Back Illuminated Scientific CMOS

Sensor Positioning Inside a Linescan Camera

Colour Reading: Chapter 6. Black body radiators

The latest trend of hybrid instrumentation

Optical Active 3D Scanning. Gianpaolo Palma

Transcription:

Efficient illuminant correction in the Local, Linear, Learned (L 3 ) method Francois G. Germain a, Iretiayo A. Akinola b, Qiyuan Tian b, Steven Lansel c, and Brian A. Wandell b,d a Center for Computer Research in Music and Acoustics, Stanford University, Stanford, CA-9435, USA; b Electrical Engineering Department, Stanford University, Stanford, CA-9435, USA; c Olympus Corporation of the Americas, Sunnyvale, CA-9485, USA; d Psychology Department, Stanford University, Stanford, CA-9435, USA. ABSTRACT To speed the development of novel camera architectures we proposed a method, L 3 (Local, Linear and Learned), that automatically creates an optimized image processing pipeline. The L 3 method assigns each sensor pixel into one of 4 classes, and applies class-dependent local linear transforms that map the sensor data from a pixel and its neighbors into the target output (e.g., CIE rendered under a D65 illuminant). The transforms are pre-computed from training data and stored in a table used for image rendering. The training data are generated by camera simulation, consisting of sensor responses and CIE rendered outputs. The sensor and rendering illuminant can be chosen equal (same-illuminant table) or different (cross-illuminant table). In the original implementation, illuminant correction is achieved with cross-illuminant tables, and one table is required for each illuminant. We find, however, that a single same-illuminant table (D65) effectively converts sensor data for many different same-illuminant conditions. Hence, we propose to render the data by applying the sameilluminant D65 table to the sensor data, followed by a linear illuminant correction transform. The mean color reproduction error using the same-illuminant table is on the order of 4 E units, which is only slightly larger than the cross-illuminant table error. This approach reduces table storage requirements significantly without substantially degrading color reproduction accuracy. Keywords: Local Linear Learned (L 3 ), image processing pipeline, illuminant correction. INTRODUCTION The L 3 method automatically generates an image processing pipeline (sensor correction, demosaicking, illuminant correction, noise reduction) for novel camera architectures. 4 The L 3 method comprises two parts: L 3 processing pipeline and L 3 training module. The L 3 processing pipeline first classifies each sensor pixel into one of many possible classes based on a statistical analysis of the responses at the pixel and its neighborhood. Then, it applies the class-associated local linear transform to map the pixel to the target output space. The L 3 training module pre-computes and stores the table of class-associated linear transforms that convert sensor data to the target CIE values. The transformed are learned from the training data using Wiener estimation. The training data consist of sensor responses and desired rendering data (i.e. CIE values for consumer photography). They are generated through simulation from a camera model that is implemented using the Image Systems Engineering Toolbox (ISET). 5,6 2. MOTIVATION: ILLUMINANT CORRECTION Illuminant correction is an essential part of the image processing pipeline that is necessary to account for the perceptual phenomenon of color constancy. 7 The visual system adapts to changes in the ambient light level and Send correspondence to Francois G. Germain. E-mail: FG (fgermain@stanford.edu), IA (iakinola@stanford.edu), QT (qytian@stanford.edu), SL (steven.lansel@olympus.com), BW (wandell@stanford.edu)

(a) (b) SENSOR DISPLA SENSOR DISPLA Fluorescent Fluorescent (Fluorescent) D65 (Target) D65 (D65) (Target) Tungsten Cross-illuminant L pipelines (I) Tungsten 3 3 (Tungsten) Same-illuminant L pipeline (SI) 3x3 linear transforms (T) Figure. Illuminant correction with L 3 image processing pipelines. (a) Cross-illuminant (I) illuminantdependent L 3 processing. (b) Same-illuminant (SI) illuminant-independent L 3 processing followed by a linear illuminant correction transform (T). color, and this adaptation has the general effect of preserving the color appearance of an object (e.g. a white shirt) across conditions (daylight to tungsten light). Because the human visual system adapts to the illuminant, to preserve color appearance the linear transforms in the L 3 method used for rendering must adapt, too. In the original algorithm, we anticipated learning a table of linear transforms for each illumination condition (cross-illuminant, Figure (a)). These cross-illuminant tables would be stored and applied for rendering. This approach pays a significant cost in computation and storage. Here, we explore an approach that separates illuminant correction from the rest of the pipeline. We learn a same-illuminant table, that renders data acquired under one illuminant into values under the same illuminant. This accomplishes demosaicking, denoising and sensor correction, but it does not perform illuminant correction (Figure (b)). We then apply an illuminant correction transform to obtain the final rendering. This architecture requires training and storing only one table of L 3 transforms to be used for all illuminants and one illuminant correction matrix for each illuminant of interest. 3. General simulation conditions 3. METHODS We used ISET 5,6 to simulate digital camera processing. The ISET camera model includes the simulation of scene radiance, optics, sensor electronics and image processing pipeline (here the L 3 pipeline). We model the camera as an f/4 lens with diffraction limited optics and focal length of 3mm. The sensor uses a RGB/W color filter array (CFA), with parameter values as in Figure 2. The L 3 algorithm uses simulated training data to learn a table of linear transforms. Each pixel in the simulated sensor data is assigned to a distinct class based on (a) its color type (R, G, B, W), the neighborhood (b) saturation pattern (no saturation, W-saturation, W&G-saturation), (c) response voltage level (2 samples spanningthevoltageswing), (d)spatialcontrast(uniformortextured). Thetransformforagivenclassisderived by optimizing a Wiener filter between the data in the pixel neighborhood in the sensor, and the corresponding ideal value in the target color space ().

Optics(diffraction-limited) Focal length(mm) 3 F-number 4 Aperture diameter (mm).75 Sensor Rows/Columns 24 24 Size (mm).4488.528 Dark noise nonuniformity (mv).4 Photoreceptor nonuniformity (% ) 2.28 3 Analog gain Analog offset (V) Sensor pixel Width/height (µm) 2.2 2.2 Fill factor.45 Dark voltage (V/sec) 5 Read noise (mv).34 Conversion gain (µv/e) 2 Voltage swing (V).8 Well capacity (electrons) 9 Food Clothes Nature Skin Hair Dupont Paint Figure 2. Optical and sensor parameter values (left) and Natural- color test chart (right) The rendering process identifies the class membership for each pixel in the sensor data and then applies the appropriate stored linear transform. The image processing pipeline, which performs demosaicking, denoising, sensor conversion comprises the application of these precomputed local, linear transforms. 3.2 L 3 tables Learning the cross-illuminant tables. Two cross-illuminant (I) tables of L 3 transforms are learned by simulating the sensor data under tungsten (I Tun ) and fluorescent (I Flu ) illumination and rendering to a target representation () under a D65 illuminant. Learning the same-illuminant tables. Same-illuminant (SI) tables of L 3 transforms are learned by simulating sensor data and the target display values using the same illuminant. We learned same-illuminant tables for D65 (SI D65 ), tungsten (SI Tun ) and fluorescent (SI Flu ). To complete the rendering, we apply an illuminant correction transform from the acquisition illuminant to the final D65 rendering. The matrix T is derived by a linear regression between the values in the acquired and target (D65) representations. T is derived from first principles and is not dependent on the camera structure or L 3. In principle, any of the SI tables can be used. In fact, it is possible to learn a same-illuminant table using sensor data from an ensemble of illuminants (e.g., a mixture of D65 and tungsten scenes). The linear transforms then reflect the properties of the multiple illuminants used in the ensemble, rather than the requirements of a specific illuminant. For example, the response level of a blue pixel under tungsten illuminant is typically quite low. This is reflected in the images obtained from the tungsten-based same-illuminant table as the L 3 processing compensates by using linear transforms with more blurring on those pixels than the D65-based same-illuminant table. We learned tables for an ensemble comprising data from a mixture of D65 and Tungsten illuminants (SI Ens ) which results in a more robust table for both illuminants. 3.3 Color evaluation To compare the I and SI methods, we used data from hyperspectral scenes and simulated test charts. Multispectral scenes. We used a small set of multispectral scenes including the Macbeth (Gretag) color chart, fruits, and people. 8 These images compare the algorithms visually.

(a) Ideal tungsten (b) Ideal D65 (c) I Tun (d) SI D65 T Figure 3. Cross-illuminant and same-illuminant algorithms perform equally well in color rendering for a tungsten illuminant. (a) The sensor image under tungsten, uncorrected. (b) The ideal image under a D65 illumination. (c) A cross-illuminant analysis for the same image but acquired under a tungsten light. The L 3 table was built from the tungsten-acquisition data and target values for a D65-rendered image. (d) A same-illuminant rendering. In this case the L 3 table was built using the same-illuminant D65 table, and the final rendering combined this table with an illuminant correction matrix T. The cross-illuminant and same-illuminant architectures produce similar images. Natural- reflectance chart. To generate a systematic and quantitative assessment of the color reproduction error, we created a Natural- test chart (Figure 2). The chart includes samples with natural object reflectances from five categories, 9 and a -sample gray color strip. The reflectances are grouped by category (2 food, 2 clothes, 2 nature, 5 hair, 5 skin and 2 paint color samples). To select the category samples we calculated 7 principal components for the entire set of natural reflectances, represented each sample by its component weights, and then formed clusters of the samples. If N samples are needed from a category, we formed N clusters for that category and chose one sample from each cluster. By selecting one sample per cluster per category we ensure that the final chart has a variety of reflectances that span the sample pool. We quantified color reproduction error using CIE2 E units. The error was the difference between the chart under a D65 illuminant and the data rendered using either the SI or I method. 4. EPERIMENTAL RESULTS We compared the quality of the color rendering using two natural hyperspectral scenes (Figures 3 and 4). Each column shows a different representation of the scene. The leftmost columns show the scene as measured under a tungsten illuminant (Figure 3(a)) and a fluorescent illuminant (Figure 4(a)). The scene rendered under a D65 illuminant, calculated from the multispectral data, is shown in panel (b). The original data rendered using the I tables is shown in panel (c), and rendered using the SI D65 table and linear transformation T (SI D65 T) is shown in (d). The I and SI D65 T renderings (panels c and d) are good approximations to the calculated D65 rendering (panel b). We quantified the color rendering quality accuracy using the Natural- test chart (Table ). The I color reproduction errors are very similar to the SI T color errors. Hence, the color accuracy of the efficient single-illuminant approach matches the cross-illuminant table for the tungsten illuminant. Interestingly, we found that the SI D65 table trained with tungsten illumination can replace the SI Tun table with only a small reduction in color reproduction accuracy (.5 E). Using a SI table derived from training on an ensemble (tungsten and D65) of illuminants reduces the color accuracy by even less (.4 E). As a result, we believe that training a SI table from data with an ensemble of illuminants may provide the best SI approach.

(a) Ideal fluorescent (b) Ideal D65 (c) I Flu (d) SI D65 T Figure 4. Cross-illuminant and same-illuminant algorithms perform equally well in color rendering for a fluorescent illuminant. (a) The sensor image under fluorescent, uncorrected. (b) The ideal image under a D65 illumination. (c) A cross-illuminant analysis for the same image but acquired under a fluorescent light. The L 3 table was built from the fluorescent-acquisition data and target values for a D65-rendered image. (d) A same-illuminant rendering. In this case the L 3 table was built using the same-illuminant D65 table, and the final rendering combined this table with an illuminant correction matrix T. The cross-illuminant and same-illuminant architectures produce similar images. A more detailed account of the color reproduction error for the cross-illuminant tables (I Tun ) and three different same-illuminant tables followed by the illuminant transform (SI Tun T, SI D65 T and SI Ens T) is presented in Figure 5 with the distribution of the E values recorded for each reflectance on the Natural- chart and the scatter plots comparingthe CIE as well as the color values of the true and reproduced colors. 5. CONCLUSIONS We compared color reproduction accuracy following illuminant correction for two versions of the L 3 algorithm. In the SI version we store a single, same-illuminant table and a set of 3 3 illuminant correction transformations. In the I version, we store a set of tables, one for each illumination. The SI method requires far less storage than the I method and costs very little in terms of color reproduction accuracy. We present different options regarding the choice of the same-illuminant table. We suggest that the best option for a same-illuminant table might be derived by training the L 3 algorithm with data simulated using an ensemble of illuminants. Color reproduction error ( E) Average (Std. dev.) 9th percentile I Tun L 3 2.5 (.8) 4.9 SI Tun L 3 with linear illuminant transform T (SI Tun T) 2.5 (2.) 5.4 SI D65 L 3 with linear illuminant transform T (SI D65 T) 4. (3.) 8.3 SI Ens L 3 with linear illuminant transform T (SI Ens T) 2.9 (2.2) 6. Table. Color reproduction errors ( E) for the Natural- color chart. The original images were simulated using a tungsten illuminant. The left column shows the training method used to derive the L 3 table. The mean reproduction error (middle) and the relatively large reproduction errors (right) are shown. The smallest reproduction error is from the cross-illuminant table. The errors for the different same-illuminant tables followed by a linear illuminant correction transform are only slightly larger.

5 Mean 2.5 5 Mean 2.5 5 Mean 4. 5 Mean 2.9 4 4 4 4 3 2 3 2 3 2 3 2 5 5 5 5 5 5 5 5 Estimated.8.6.4.2.5 True Estimated.8.6.4.2.5 True Estimated.8.6.4.2.5 True Estimated.8.6.4.2.5 True 5 5 5 5 (a) I Tun 5 5 5 5 (b) SI Tun T 5 5 5 5 (c) SI D65 T 5 5 5 5 (d) SI Ens T Figure 5. Color error distributions for a Natural- chart acquired under a tungsten illuminant and rendered using different L 3 tables. Histograms of the distribution of color reproduction errors in E units (top), scatter plots between true and estimated coordinates (middle) and scatter plots between true and estimated coordinates (bottom) for the chart reflectances for different L 3 tables: (a) the cross-illuminant table (I Tun), (b) the same-illuminant table trained under tungsten followed by the illuminant transform T (SI Tun T), (c) the same-illuminant table trained under D65 followed by the illuminant transform T (SI D65 T) and (d) the same-illuminant table trained under an ensemble of illuminants followed by the illuminant transform T (SI Ens T). REFERENCES [] Tian, Q., Lansel, S., Farrell, J. E., and Wandell, B. A., Automating the design of image processing pipelines for novel color filter arrays: Local, Linear, Learned (L 3 ) method, in [IS&T/SPIE Electronic Imaging], 923K 923K, International Society for Optics and Photonics (24). [2] Lansel, S. P. and Wandell, B. A., Local linear learned image processing pipeline, in [Imaging Systems and Applications], Optical Society of America (2). [3] Lansel, S. P., Local Linear Learned Method for Image and Reflectance Estimation, PhD thesis, Stanford University (2). [4] Lansel, S. P. and Wandell, B. A., Learning of image processing pipeline for digital imaging devices, (Dec. 7 22). [5] Farrell, J. E., iao, F., Catrysse, P. B., and Wandell, B. A., A simulation tool for evaluating digital camera image quality, in[electronic Imaging 24], 24 3, International Society for Optics and Photonics(23).

[6] Farrell, J. E., Catrysse, P. B., and Wandell, B. A., Digital camera simulation, Applied Optics 5(4), A8 A9 (22). [7] Wandell, B. A., [Foundations of Vision], Sinauer Associates (995). [8] http://imageval.com/scene-database/. [9] Vrhel, M. J., Gershon, R., and Iwan, L. S., Measurement and analysis of object reflectance spectra, Color Research & Application 9(), 4 9 (994). [] Sharma, G., Wu, W., and Dalal, E. N., The CIEDE2 color-difference formula: Implementation notes, supplementary test data, and mathematical observations, Color Research & Application 3(), 2 3 (25).