Optimal Denoising of Natural Images and their Multiscale Geometry and Density

Similar documents
Denoising an Image by Denoising its Components in a Moving Frame

IMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING

Learning how to combine internal and external denoising methods

IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING

Sparse & Redundant Representations and Their Applications in Signal and Image Processing

Markov Random Fields and Gibbs Sampling for Image Denoising

Locally Adaptive Regression Kernels with (many) Applications

IMAGE DENOISING TO ESTIMATE THE GRADIENT HISTOGRAM PRESERVATION USING VARIOUS ALGORITHMS

Adaptive Kernel Regression for Image Processing and Reconstruction

Locally Adaptive Learning for Translation-Variant MRF Image Priors

Image Restoration Using DNN

Blind Image Deblurring Using Dark Channel Prior

Introduction to patch-based approaches for image processing F. Tupin

Image Denoising using Locally Learned Dictionaries

SUPPLEMENTARY MATERIAL

Statistical image models

Computational Photography Denoising

Generalized Tree-Based Wavelet Transform and Applications to Patch-Based Image Processing

CLASS-SPECIFIC POISSON DENOISING BY PATCH-BASED IMPORTANCE SAMPLING

Patch-Based Color Image Denoising using efficient Pixel-Wise Weighting Techniques

Implementation of the Non-Local Bayes (NL-Bayes) Image Denoising Algorithm

Locally Adaptive Regression Kernels with (many) Applications

Explore the Power of External Data in Denoising Task

One Network to Solve Them All Solving Linear Inverse Problems using Deep Projection Models

IMAGE DENOISING BY TARGETED EXTERNAL DATABASES

Graphical Models for Computer Vision

arxiv: v1 [cs.cv] 2 Dec 2015

Available online at ScienceDirect. Procedia Computer Science 54 (2015 ) Mayank Tiwari and Bhupendra Gupta

Improved Non-Local Means Algorithm Based on Dimensionality Reduction

A non-local algorithm for image denoising

Image pyramids and their applications Bill Freeman and Fredo Durand Feb. 28, 2006

An Optimized Pixel-Wise Weighting Approach For Patch-Based Image Denoising

08 An Introduction to Dense Continuous Robotic Mapping

Expected Patch Log Likelihood with a Sparse Prior

Filters. Advanced and Special Topics: Filters. Filters

A Comparative Study & Analysis of Image Restoration by Non Blind Technique

AN ALGORITHM FOR BLIND RESTORATION OF BLURRED AND NOISY IMAGES

Instance-based Learning CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2015

Multi-frame super-resolution with no explicit motion estimation

Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis

When Sparsity Meets Low-Rankness: Transform Learning With Non-Local Low-Rank Constraint For Image Restoration

Perception Viewed as an Inverse Problem

Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling

Image Redundancy and Non-Parametric Estimation for Image Representation

Synthesis and Analysis Sparse Representation Models for Image Restoration. Shuhang Gu 顾舒航. Dept. of Computing The Hong Kong Polytechnic University

Image Processing with Nonparametric Neighborhood Statistics

IMAGE DENOISING BY TARGETED EXTERNAL DATABASES

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

Image Restoration with Deep Generative Models

GRID WARPING IN TOTAL VARIATION IMAGE ENHANCEMENT METHODS. Andrey Nasonov, and Andrey Krylov

Use of Extreme Value Statistics in Modeling Biometric Systems

The Trainable Alternating Gradient Shrinkage method

Machine Learning / Jan 27, 2010

Bilevel Sparse Coding

Spatio-Temporal Markov Random Field for Video Denoising

A Fast Non-Local Image Denoising Algorithm

CHAPTER VIII SEGMENTATION USING REGION GROWING AND THRESHOLDING ALGORITHM

Scene Grammars, Factor Graphs, and Belief Propagation

Analysis of Various Issues in Non-Local Means Image Denoising Algorithm

Investigating Image Inpainting via the Alternating Direction Method of Multipliers

Face Hallucination Based on Eigentransformation Learning

Image Processing using Smooth Ordering of its Patches

MULTIFRAME RAW-DATA DENOISING BASED ON BLOCK-MATCHING AND 3-D D FILTERING FOR LOW-LIGHT LIGHT IMAGING AND STABILIZATION

Blind Deblurring using Internal Patch Recurrence. Tomer Michaeli & Michal Irani Weizmann Institute

Exploring Curve Fitting for Fingers in Egocentric Images

Scene Grammars, Factor Graphs, and Belief Propagation

K-means and Hierarchical Clustering

Note Set 4: Finite Mixture Models and the EM Algorithm

Bayesian Methods in Vision: MAP Estimation, MRFs, Optimization

Image Restoration and Background Separation Using Sparse Representation Framework

Using the Forest to See the Trees: Context-based Object Recognition

Image Denoising Based on Hybrid Fourier and Neighborhood Wavelet Coefficients Jun Cheng, Songli Lei

Convex Optimization MLSS 2015

Large Scale Data Analysis Using Deep Learning

BM3D Image Denoising using Learning-based Adaptive Hard Thresholding

Missing Data Analysis for the Employee Dataset

SEMI-BLIND IMAGE RESTORATION USING A LOCAL NEURAL APPROACH

External Patch Prior Guided Internal Clustering for Image Denoising

How does bilateral filter relates with other methods?

Detecting Salient Contours Using Orientation Energy Distribution. Part I: Thresholding Based on. Response Distribution

DS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University

Clustering: Classic Methods and Modern Views

Hybrid Quasi-Monte Carlo Method for the Simulation of State Space Models

IMA Preprint Series # 2052

arxiv: v1 [cs.cv] 18 Jan 2019

Divide and Conquer Kernel Ridge Regression

arxiv: v1 [cs.cv] 30 Oct 2018

Image Denoising AGAIN!?

Image Processing Via Pixel Permutations

Image Restoration Using Gaussian Mixture Models With Spatially Constrained Patch Clustering

DENOISING AN IMAGE BY DENOISING ITS CURVATURE IMAGE. Marcelo Bertalmío and Stacey Levine. IMA Preprint Series #2411.

Image Denoising via Group Sparse Eigenvectors of Graph Laplacian

Chapter 3. Bootstrap. 3.1 Introduction. 3.2 The general idea

Robert Collins CSE598G. Robert Collins CSE598G

All images are degraded

Digital Image Processing Laboratory: MAP Image Restoration

ICA mixture models for image processing

Learning to Perceive Transparency from the Statistics of Natural Scenes

Structural Similarity Optimized Wiener Filter: A Way to Fight Image Noise

Multi-scale Statistical Image Models and Denoising

Transcription:

Optimal Denoising of Natural Images and their Multiscale Geometry and Density Department of Computer Science and Applied Mathematics Weizmann Institute of Science, Israel. Joint work with Anat Levin (WIS), Fredo Durand and Bill Freeman (MIT). June 2013 [Levin and Nadler, CVPR 2011] [Levin, Nadler, Durand, Freeman, ECCV 2012]

Classical Statistical Estimation vs. Image Restoration Classical Statistical Inference: {x i } n i=1 are n i.i.d. random samples (observations) from some unknown density p(x) (in parametric setting, p(x) = p θ (x) ). Task: Given {x i } wish to estimate p(x) (or θ).

Classical Statistical Estimation vs. Image Restoration Classical Statistical Inference: {x i } n i=1 are n i.i.d. random samples (observations) from some unknown density p(x) (in parametric setting, p(x) = p θ (x) ). Task: Given {x i } wish to estimate p(x) (or θ). Key result: Consistency As n, ˆp(x) p(x), and ˆθ θ

Restoration Tasks and Image Priors Many image restoration tasks: denoising, super-resolution, deblurring, etc, are different:

Restoration Tasks and Image Priors Many image restoration tasks: denoising, super-resolution, deblurring, etc, are different: Have only one (or even less) noisy observation per each unknown.

Restoration Tasks and Image Priors Many image restoration tasks: denoising, super-resolution, deblurring, etc, are different: Have only one (or even less) noisy observation per each unknown. Image Restoration Tasks are ill-posed problems.

Restoration Tasks and Image Priors Many image restoration tasks: denoising, super-resolution, deblurring, etc, are different: Have only one (or even less) noisy observation per each unknown. Image Restoration Tasks are ill-posed problems. Given a degraded image y, there are multiple valid explanations x.

Restoration Tasks and Image Priors Many image restoration tasks: denoising, super-resolution, deblurring, etc, are different: Have only one (or even less) noisy observation per each unknown. Image Restoration Tasks are ill-posed problems. Given a degraded image y, there are multiple valid explanations x. Natural image priors are crucial to rule out unlikely solutions.

On the Strength of Natural Image Priors

On the Strength of Natural Image Priors Natural k k patches are extremely sparse in the set of 256 k k patches

Optimal Image Restoration Practical / Computational issues limit most current image priors to local patches.

Optimal Image Restoration Practical / Computational issues limit most current image priors to local patches. Hence, the restoration results are suboptimal.

Towards Optimal Image Restoration

Towards Optimal Image Restoration Fundamental Theoretical Questions:

Towards Optimal Image Restoration Fundamental Theoretical Questions: - How accurate are currently used priors?

Towards Optimal Image Restoration Fundamental Theoretical Questions: - How accurate are currently used priors? - How much gain is possible by increasing patch size?

Towards Optimal Image Restoration Fundamental Theoretical Questions: - How accurate are currently used priors? - How much gain is possible by increasing patch size? - Computational issues aside, what is the optimal possible restoration? Does it have a zero error?

Towards Optimal Image Restoration Fundamental Theoretical Questions: - How accurate are currently used priors? - How much gain is possible by increasing patch size? - Computational issues aside, what is the optimal possible restoration? Does it have a zero error? - How do these questions relate to natural image statistics?

Towards Optimal Image Restoration Fundamental Theoretical Questions: - How accurate are currently used priors? - How much gain is possible by increasing patch size? - Computational issues aside, what is the optimal possible restoration? Does it have a zero error? - How do these questions relate to natural image statistics? This talk will try to address some of these issues for a specific task: denoising

Outline Denoising of Natural Images: - Ill posed problem - solved using priors - Optimal prior = (unknown) density of natural images

Outline Denoising of Natural Images: - Ill posed problem - solved using priors - Optimal prior = (unknown) density of natural images Questions:

Outline Denoising of Natural Images: - Ill posed problem - solved using priors - Optimal prior = (unknown) density of natural images Questions: i) What is the optimal achievable denoising? How is it related to the multiscale geometry and density of natural images?

Outline Denoising of Natural Images: - Ill posed problem - solved using priors - Optimal prior = (unknown) density of natural images Questions: i) What is the optimal achievable denoising? How is it related to the multiscale geometry and density of natural images? ii) How far from optimal are current algorithms with approximate priors?

Outline Denoising of Natural Images: - Ill posed problem - solved using priors - Optimal prior = (unknown) density of natural images Questions: i) What is the optimal achievable denoising? How is it related to the multiscale geometry and density of natural images? ii) How far from optimal are current algorithms with approximate priors? iii) How can we further improve current algorithms?

Image Denoising - The Final Frontier [Is denoising dead? / Chatterjee & Milanfar]

Image Denoising - The Final Frontier [Is denoising dead? / Chatterjee & Milanfar] Should you do your Ph.D. in image denoising?

Image Denoising - The Final Frontier [Is denoising dead? / Chatterjee & Milanfar] Should you do your Ph.D. in image denoising? Answer at end of talk...

Natural Image Denoising Problem Setting: x : unknown noise free natural image y = x + n : observed image corrupted by noise. In this talk, n - additive iid Gaussian N(0, σ 2 ).

Natural Image Denoising Problem Setting: x : unknown noise free natural image y = x + n : observed image corrupted by noise. In this talk, n - additive iid Gaussian N(0, σ 2 ). Task: Given y, construct clean version ŷ (estimate of x).

Natural Image Denoising Problem Setting: x : unknown noise free natural image y = x + n : observed image corrupted by noise. In this talk, n - additive iid Gaussian N(0, σ 2 ). Task: Given y, construct clean version ŷ (estimate of x). Performance Measure: MSE = E[ x ŷ 2 ]

Natural Image Denoising Problem Setting: x : unknown noise free natural image y = x + n : observed image corrupted by noise. In this talk, n - additive iid Gaussian N(0, σ 2 ). Task: Given y, construct clean version ŷ (estimate of x). Performance Measure: MSE = E[ x ŷ 2 ] Challenge: Ill posed problem: #unknowns = #observations

Denoising Algorithms Priors: All denoising algorithms employ implicitly or explicitly some prior on the unknown (noise-free) image. Many different priors and denoising algorithms. 1980 s - Gabor filters, anisotropic diffusion 1990 s - wavelet based methods, total variation 2000 s - sparse representations,

Denoising Algorithms Priors: All denoising algorithms employ implicitly or explicitly some prior on the unknown (noise-free) image. Many different priors and denoising algorithms. 1980 s - Gabor filters, anisotropic diffusion 1990 s - wavelet based methods, total variation 2000 s - sparse representations, 05 Non-Local Means Algorithm

The NL-Means Algorithm L. Yaroslavsky already suggested similar ideas in 1985. [Buades & al 2005] [Awate & Whittaker 2006] [Barash & Comaniciu 2004] where ŷ(x i ) = 1 D(x i ) D(x i ) = pixels j pixels j K ε (y(x i ), y(x j ))y(x j ) K ε (y(x i ), y(x j ))

The NL-Means Algorithm L. Yaroslavsky already suggested similar ideas in 1985. [Buades & al 2005] [Awate & Whittaker 2006] [Barash & Comaniciu 2004] where In matrix form ŷ(x i ) = 1 D(x i ) D(x i ) = pixels j pixels j K ε (y(x i ), y(x j ))y(x j ) K ε (y(x i ), y(x j )) ŷ = D 1 Wy One interpretation: Random walk on image patches. [Singer, Shkolnisky, N. 08 ]

Example: Input Image

Example: Original noise-free image

Example: Result of BM3D

Example: Result of KSVD

Image Denoising, Smoothness, Sparsity, Priors Key Questions About Natural Image Denoising:

Image Denoising, Smoothness, Sparsity, Priors Key Questions About Natural Image Denoising: - Optimal Denoising in this setup? Relation to geometry and density of natural image patches?

Image Denoising, Smoothness, Sparsity, Priors Key Questions About Natural Image Denoising: - Optimal Denoising in this setup? Relation to geometry and density of natural image patches? - Are we there yet? How much better can we improve on BM3D or other denoising algorithms?

Statistical Framework First consider denoising algorithms that estimate the central pixel x c from a k k window around it, dimension d = k 2. In other words, think of x, y as both k k patches. Goal: Given y = x + n, estimate central pixel of x as best as possible. Quality Measure: Mean Squared Error (MSE) MSE = E[(ŷ c x c ) 2 ], PSNR = 10 log 10 MSE

Optimal Natural Image Denoising The Bayesian Minimum Mean Squared Error Estimator µ(y) = E[x c y] = p(x c y)x c dx = p(y x)xc dx pd (x)p(y x)dx where p d (x) - density of d = k k patches of natural images.

Optimal Natural Image Denoising The Bayesian Minimum Mean Squared Error Estimator µ(y) = E[x c y] = p(x c y)x c dx = p(y x)xc dx pd (x)p(y x)dx where p d (x) - density of d = k k patches of natural images. NL-means ŷ(x i ) = j K(y(x j), y(x i ))y c (x j ) j K(y(x i), y(x j ))

Optimal Natural Image Denoising The Bayesian Minimum Mean Squared Error Estimator µ(y) = E[x c y] = p(x c y)x c dx = p(y x)xc dx pd (x)p(y x)dx where p d (x) - density of d = k k patches of natural images. NL-means ŷ(x i ) = j K(y(x j), y(x i ))y c (x j ) j K(y(x i), y(x j )) can be viewed as non-parametric approximation of the MMSE estimator using noisy image patches.

Upper and Lower Bounds on MMSE If p d (x) were known - in principle could compute MMSE and see how far we are from optimality (at fixed window size d).

Upper and Lower Bounds on MMSE If p d (x) were known - in principle could compute MMSE and see how far we are from optimality (at fixed window size d). Challenge: Derive lower bound on MMSE d w/out knowing p d (x)?

Upper and Lower Bounds on MMSE If p d (x) were known - in principle could compute MMSE and see how far we are from optimality (at fixed window size d). Challenge: Derive lower bound on MMSE d w/out knowing p d (x)? Solution: A trick...

A brute force / elegant non-parametric approach Don t know p d (x), but can sample from it.

A brute force / elegant non-parametric approach Don t know p d (x), but can sample from it. Approximate two identical yet different representations of MSE.

A brute force / elegant non-parametric approach Don t know p d (x), but can sample from it. Approximate two identical yet different representations of MSE. ERROR representation MSE = p(x) p(y x)(ˆµ(y) x c ) 2 dydx

A brute force / elegant non-parametric approach Don t know p d (x), but can sample from it. Approximate two identical yet different representations of MSE. ERROR representation MSE = p(x) p(y x)(ˆµ(y) x c ) 2 dydx VARIANCE representation MSE = p σ (y) p(x y)(ˆµ(y) x c ) 2 dxdy

A brute force / elegant non-parametric approach Don t know p d (x), but can sample from it.

A brute force / elegant non-parametric approach Don t know p d (x), but can sample from it. Consider a huge set of N = 10 10 natural image patches. Approximate MMSE d non-parametrically. where for Gaussian noise and d = k 2. ˆµ(y) = p(y x) = 1 N 1 N i p(y x i)x i,c i p(y x, (1) i) 1 x y 2 (2πσ 2 e 2σ ) d/2 2 (2)

Upper and Lower Bound on MMSE Given a set of M noisy pairs {( x j, y j )} M j=1 and another independent set of N clean patches {x i } N i=1, both randomly sampled from natural images, we compute MMSE U = 1 (ˆµ(y j ) x j,c ) 2 (3) M j MMSE L = 1 ˆV(y j ) (4) M where ˆV(y j ) is the approximated variance: j ˆV(y j ) = 1 N i p(y j x i )(ˆµ(y j ) x i,c ) 2 1 N i p(y j x i ) (5)

Upper and Lower Bound on MMSE Given a set of M noisy pairs {( x j, y j )} M j=1 and another independent set of N clean patches {x i } N i=1, both randomly sampled from natural images, we compute MMSE U = 1 (ˆµ(y j ) x j,c ) 2 (3) M j MMSE L = 1 ˆV(y j ) (4) M where ˆV(y j ) is the approximated variance: j ˆV(y j ) = 1 N i p(y j x i )(ˆµ(y j ) x i,c ) 2 1 N i p(y j x i ) (5) Key Difference: MMSE U knows true noise-free patch x, MMSE L does not know x,

The Theoretical Part MMSE U is the MSE of yet another (not very fast) denoising algorithm. It is an upper bound on the optimal MSE.

The Theoretical Part MMSE U is the MSE of yet another (not very fast) denoising algorithm. It is an upper bound on the optimal MSE. Claim: In expectation, MMSE L is a lower bound on the MSE.

The Theoretical Part MMSE U is the MSE of yet another (not very fast) denoising algorithm. It is an upper bound on the optimal MSE. Claim: In expectation, MMSE L is a lower bound on the MSE. If N 1 is sufficiently large so that MMSE U MMSE L - then we have an accurate estimate of the MMSE!

The Theoretical Part Claim with ( ) 1 E N [ ˆV(y)] = V(y) + C(y)B(y) + o, (6) N B(y) = E[x 2 c y] 3E[x c y] 2 2E σ [x 2 c y] +4E[x c y]e σ [x c y] (7) C(y) = 1 p σ (y) N (4πσ 2 ) d/2 p(y) 2 (8) where p σ and E σ [ ] denote probability and expectation of random variables whose noise standard deviation is reduced from σ to σ = σ/ 2.

Theory Part II Next, we consider a local Gaussian approximation for p(x) around x = y, e.g., a Laplace approximation consisting of a second order Taylor expansion of ln p(x). For a Gaussian distribution analytic computation of bias Claim For a Gaussian distribution, B(y) 0 for all y.

Numerical Example 2 0 x 10 3 B(y j ) 2 4 6 8 0 500 1000 1500 2000 j

Experiment Design Set of N = 10 10 natural image patches out of 20K images from LabelMe dataset. Independent set of M = 2000 patches from same dataset x. Add noise at different levels, σ = 18, 55, 170, and with different window sizes, k = 3, 4,..., 20. Compared MMSE L, MMSE U as well as MSE of various state-of-the-art denoising algorithms.

Experiments: σ = 18 PSNR 35 34 33 32 MMSE U MMSE L LMMSE BM3D KSVD GSM 31 30 0 5 10 15 20 25 Support size k

MMSE d vs window size 30 29 PSNR 28 27 26 25 4 6 8 10 12 patch size For small patches / large noise non-parametric approach can accurately estimate MMSE. [Levin and N. CVPR 2011]

MMSE d vs window size 30 29 PSNR 28 27 26 25 4 6 8 10 12 patch size Extrapolation: What happens as d?

Towards Results of Levin and Nadler [CVPR 2011] imply that with small patches current algorithms are close to optimal.

Towards Results of Levin and Nadler [CVPR 2011] imply that with small patches current algorithms are close to optimal. Open Questions:

Towards Results of Levin and Nadler [CVPR 2011] imply that with small patches current algorithms are close to optimal. Open Questions: -How much gain is possible by increasing patch size?

Towards Results of Levin and Nadler [CVPR 2011] imply that with small patches current algorithms are close to optimal. Open Questions: -How much gain is possible by increasing patch size? - Computational issues aside, what is the optimal possible restoration? Does it have a zero error?

Patch Complexity vs. PSNR gain Empirical Results: Assign each pixel to group G d where d is largest window size with reliable estimation.

Patch Complexity vs. PSNR gain Empirical Results: Assign each pixel to group G d where d is largest window size with reliable estimation. PSNR vs. number of pixels in window, for different groups G d 35 G 7 30 PSNR 30 G 11 G 14 G 17 G 21 G 25 G 30 PSNR 25 G 40 G 317 0 15 10 20 30 40 0 50 100 Number of pixelsl Number of pixelsl 150 1D signals, σ = 10 2D signals, σ = 20 25 20 G 49 G 69 G 113 G 149 G 197

Patch Complexity vs. PSNR gain Higher curves correspond to smooth regions, which flatten at larger patch dimensions, require smaller external database. Lower curves correspond to textured regions, which not only run out of samples sooner, but also their curves flatten earlier, and require huge external database.

Patch Complexity vs. PSNR gain Key conclusion: Law of Marginal Return

Patch Complexity vs. PSNR gain Key conclusion: Law of Marginal Return When increasing window size, Smooth regions: large gain, no need to increase sample size

Patch Complexity vs. PSNR gain Key conclusion: Law of Marginal Return When increasing window size, Smooth regions: large gain, no need to increase sample size Texture / edges: small gain, large increase in sample size

Adaptive Non-parametric Denoising For each pixel x i, find largest window w d for which ˆµ d is still reliable.

Adaptive Non-parametric Denoising For each pixel x i, find largest window w d for which ˆµ d is still reliable. σ 20 35 50 75 100 Optimal Fixed 32.4 30.1 28.7 27.2 26.0 Adaptive 33.0 30.5 29.0 27.5 26.4 BM3D 33.2 30.3 28.6 26.9 25.6

Adaptive Non-parametric Denoising For each pixel x i, find largest window w d for which ˆµ d is still reliable. σ 20 35 50 75 100 Optimal Fixed 32.4 30.1 28.7 27.2 26.0 Adaptive 33.0 30.5 29.0 27.5 26.4 BM3D 33.2 30.3 28.6 26.9 25.6 Similar idea suggested for adaptive Non-Local Means and Boulanger. 2006]. [Kervrann Challenge: Develop Practical Algorithm Based on this Adaptivity Principle

PSNR vs. window size What is convergence rate of MMSE d to MMSE?

PSNR vs. window size What is convergence rate of MMSE d to MMSE? 28 26 PSNR 24 22 Data Polynomial fit Exponential fit 0 50 100 d

PSNR vs. window size Empirically: MMSE d = e + c d α with α 1

PSNR vs. window size Empirically: MMSE d = e + c d α with α 1 Can this be predicted theoretically?

Dead Leaves Model Simple Image Formation Model: Dead Leaves [Matheron 68 ]

Dead Leaves Model Simple Image Formation Model: Dead Leaves [Matheron 68 ] Image = random collection of finite size piece-wise constant regions Region intensity = random variable with uniform distribution.

Multiscale Geometry of Natural Images Scale Invariance: many natural image statistics are scale invariant: Down-sampling natural images does not change gradient distribution, segment sizes, etc. [Ruderman 97, Field 78, etc.]

Multiscale Geometry of Natural Images Scale Invariance: many natural image statistics are scale invariant: Down-sampling natural images does not change gradient distribution, segment sizes, etc. [Ruderman 97, Field 78, etc.] Claim Under scale invariance Pr[pixel belongs to region of size s pixels ] 1/s [Alvarez, Gousseau, Morel, 99 ]

Scale Invariance and the dead leaves model Our Key Contribution: Claim Dead leaves model with scale invariance (and edge oracle) implies strictly positive MMSE and power law convergence with α = 1 MMSE d = e + c d α

Scale Invariance and the dead leaves model Our Key Contribution: Claim Dead leaves model with scale invariance (and edge oracle) implies strictly positive MMSE and power law convergence MMSE d = e + c d α with α = 1 Note that e = MMSE so we can extrapolate empirical curve and estimate PSNR.

Extrapolated PSNR vs. Current Algorithms σ 35 50 75 100 Extrapolated bound 30.6 28.8 27.3 26.3 KSVD 28.7 26.9 25.0 23.7 BM3D 30.0 28.1 26.3 25.0 EPLL 29.8 28.1 26.3 25.1

Extrapolated PSNR vs. Current Algorithms σ 35 50 75 100 Extrapolated bound 30.6 28.8 27.3 26.3 KSVD 28.7 26.9 25.0 23.7 BM3D 30.0 28.1 26.3 25.0 EPLL 29.8 28.1 26.3 25.1 Still some modest room for improvement

Summary - Understanding the relation between natural image statistics and restoration taks is a fundamental problem.

Summary - Understanding the relation between natural image statistics and restoration taks is a fundamental problem. - Statistical framework to study optimal natural image denoising.

Summary - Understanding the relation between natural image statistics and restoration taks is a fundamental problem. - Statistical framework to study optimal natural image denoising. - Non-parametric methods - law of diminishing return.

Summary - Understanding the relation between natural image statistics and restoration taks is a fundamental problem. - Statistical framework to study optimal natural image denoising. - Non-parametric methods - law of diminishing return. - most further gain in smooth regions adaptive approach.

Summary - Understanding the relation between natural image statistics and restoration taks is a fundamental problem. - Statistical framework to study optimal natural image denoising. - Non-parametric methods - law of diminishing return. - most further gain in smooth regions adaptive approach. - Natural image statistics pose inherent limits on denoising. Empirically, power law convergence vs. window size, supported by Dead Leaves + Scale Invariance model.

Summary - Understanding the relation between natural image statistics and restoration taks is a fundamental problem. - Statistical framework to study optimal natural image denoising. - Non-parametric methods - law of diminishing return. - most further gain in smooth regions adaptive approach. - Natural image statistics pose inherent limits on denoising. Empirically, power law convergence vs. window size, supported by Dead Leaves + Scale Invariance model. - Still room for improvement (not by non-parametric method).

Summary - Understanding the relation between natural image statistics and restoration taks is a fundamental problem. - Statistical framework to study optimal natural image denoising. - Non-parametric methods - law of diminishing return. - most further gain in smooth regions adaptive approach. - Natural image statistics pose inherent limits on denoising. Empirically, power law convergence vs. window size, supported by Dead Leaves + Scale Invariance model. - Still room for improvement (not by non-parametric method). - Implications to other low level vision tasks - deconvolution, super-resolution.

Summary - Understanding the relation between natural image statistics and restoration taks is a fundamental problem. - Statistical framework to study optimal natural image denoising. - Non-parametric methods - law of diminishing return. - most further gain in smooth regions adaptive approach. - Natural image statistics pose inherent limits on denoising. Empirically, power law convergence vs. window size, supported by Dead Leaves + Scale Invariance model. - Still room for improvement (not by non-parametric method). - Implications to other low level vision tasks - deconvolution, super-resolution. THE END / THANK YOU!