Presentation is loading. Please wait.

Presentation is loading. Please wait.

(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.

Similar presentations


Presentation on theme: "(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation."— Presentation transcript:

1 (1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation (shown zero mean for notational convenience) : If we rotate coordinates to the Fourier basis, the covariance matrix in that basis will be diagonal. So in that model, each Fourier transform coefficient is an independent Gaussian random variable of covariance Image pixels Inverse covariance matrix

2 Power spectra of typical images Experimentally, the power spectrum as a function of Fourier frequency is observed to follow a power law. http://www.cns.nyu.edu/pub/eero/simoncelli05a-preprint.pdf

3 Random draw from Gaussian spectral model http://www.cns.nyu.edu/pub/eero/simoncelli05a-preprint.pdf

4 Noise removal (in frequency domain), under Gaussian assumption Variance of white, Gaussian additive noise Observed Fourier component Estimated Fourier component Power law prior probability on estimated Fourier component Setting to zero the derivative of the the log probability of X gives an analytic form for the optimal estimate of X (or just complete the square): Posterior probability for X

5 Noise removal, under Gaussian assumption original With Gaussian noise of std. dev. 21.4 added, giving PSNR=22.06 (1) Denoised with Gaussian model, PSNR=27.87 http://www.cns.nyu.edu/pub/eero/simoncelli05a-preprint.pdf (try to ignore JPEG compression artifacts from the PDF file)

6 (2) The wavelet marginal model Histogram of wavelet coefficients, c, for various images. Parameter determining width of distribution Parameter determining peakiness of distribution Wavelet coefficient value http://www.cns.nyu.edu/pub/eero/simoncelli05a-preprint.pdf

7 Random draw from the wavelet marginal model http://www.cns.nyu.edu/pub/eero/simoncelli05a-preprint.pdf

8 And again something that is reminiscent of operations found in V1…

9 An application of image pyramids: noise removal

10 Image statistics (or, mathematically, how can you tell image from noise?) Noisy image

11 Clean image

12 Pixel representation, image histogram

13 Pixel representation, noisy image histogram

14 bandpassed representation image histogram

15 Pixel domain noise image and histogram

16 Bandpass domain noise image and histogram

17 Noise-corrupted full-freq and bandpass images But want the bandpass image histogram to look like this

18 P(x, y) = P(x|y) P(y) so P(x|y) P(y) = P(y|x) P(x) P(x, y) = P(x|y) P(y) so P(x|y) P(y) = P(y|x) P(x) and P(x|y) = P(y|x) P(x) / P(y) Bayes theorem The parameters you want to estimate What you observe Prior probability Likelihood function Constant w.r.t. parameters x. P(x, y) = P(x|y) P(y) By definition of conditional probability Using that twice

19 P(x) Bayesian MAP estimator for clean bandpass coefficient values Let x = bandpassed image value before adding noise. Let y = noise-corrupted observation. By Bayes theorem P(x|y) = k P(y|x) P(x) P(y|x) P(x|y) P(y|x) y y = 25

20 Bayesian MAP estimator Let x = bandpassed image value before adding noise. Let y = noise-corrupted observation. By Bayes theorem P(x|y) = k P(y|x) P(x) y P(y|x) P(x|y) y = 50

21 Bayesian MAP estimator Let x = bandpassed image value before adding noise. Let y = noise-corrupted observation. By Bayes theorem P(x|y) = k P(y|x) P(x) y P(y|x) P(x|y) y = 115

22 P(x) P(y|x) y y = 25 P(x|y) y P(y|x) P(x|y) y = 115 For small y: probably it is due to noise and y should be set to 0 For large y: probably it is due to an image edge and it should be kept untouched

23 MAP estimate,, as function of observed coefficient value, y http://www-bcs.mit.edu/people/adelson/pub_pdfs/simoncelli_noise.pdf Simoncelli and Adelson, Noise Removal via Bayesian Wavelet Coring

24 original With Gaussian noise of std. dev. 21.4 added, giving PSNR=22.06 (1) Denoised with Gaussian model, PSNR=27.87 (2) Denoised with wavelet marginal model, PSNR=29.24 http://www.cns.nyu.edu/pub/eero/simoncelli05a-preprint.pdf

25 M. F. Tappen, B. C. Russell, and W. T. Freeman, Efficient graphical models for processing images IEEE Conf. on Computer Vision and Pattern Recognition (CVPR) Washington, DC, 2004.

26 Motivation for wavelet joint models Note correlations between the amplitudes of each wavelet subband. http://www.cns.nyu.edu/pub/eero/simoncelli05a-preprint.pdf

27 Statistics of pairs of wavelet coefficients Contour plots of the joint histogram of various wavelet coefficient pairs Conditional distributions of the corresponding wavelet pairs http://www.cns.nyu.edu/pub/eero/simoncelli05a-preprint.pdf

28 (3) Gaussian scale mixtures Wavelet coefficient probability A mixture of Gaussians of scaled covariances observed Gaussian scale mixture model simulation z is a spatially varying hidden variable that can be used to (a) Create the non-gaussian histograms from a mixture of Gaussian densities, and (b) model correlations between the neighboring wavelet coefficients.

29 original With Gaussian noise of std. dev. 21.4 added, giving PSNR=22.06 (1) Denoised with Gaussian model, PSNR=27.87 http://www.cns.nyu.edu/pub/eero/simoncelli05a-preprint.pdf (3) Denoised with Gaussian scale mixture model, PSNR=30.86 (2) Denoised with wavelet marginal model, PSNR=29.24


Download ppt "(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation."

Similar presentations


Ads by Google