Download presentation

Presentation is loading. Please wait.

Published byLuis Bachus Modified over 3 years ago

1
Probabilistic Inference and Learning in Computer Vision An extract from the BMVC2000 pre-conference tutorial given by: Prof. Andrew Blake Microsoft Research

2
Learning low-level vision, Freeman and Pasztor, Proc. ICCV99. This paper proposes a persuasive general approach to inference in image arrays. The classic application is restoration of degraded images, including super-resolution. This is a classic Bayesian piece of work, the latest in an honourable succession that began with “intrinsic images” (Barrow and Tanenbaum 1978,) and moved on the regularisation (Poggio et al. 1983,) via Markov random fields (MRF) and Gibbs sampling (Geman and Geman 1984,) and probabilistic graphical models (Pearl 1988.) It characterises the striking new trend towards exemplar-based learning. It’s certainly bracing stuff- where’s the catch? Learning graphical models of images, videos and their spatial transformations, Frey and Jojic, Proc. UAI2000. They have put together an exciting story that uses “latent variable modelling,” second nature in the probabilistic inference (NIPS) community, to explain and analyse images and image sequences. The exciting part is that, apparently, all you have do is describe how an image is constructed, and you automatically get an analysis of the images. The trick is, you just take the same description and push it through EM machine. It seems almost miraculous, in the same way that declarative programming (PROLOG) seems miraculous, that the analytical machinery is generated for you automatically. Is there a catch here, or should we all be doing this?

3
Probabilistic Graphical Models for image motion analysis (Frey and Jojic, 99/00) x z c z Latent image model. x is the unknown (or latent) image. z is the image produced by the model, or found in real life. e.g. p(z|x)=N(x, ) Mixture Model. c is the unknown cluster centre. z is the sampled value. e.g. p(z|c)=N( c, c ) z c Continuous random variableDiscrete random variable

4
z Transformed latent image model. P(l=L) = l, p(z|x,l)=N(T l x, ) Principle Components/Factor Analysis. p(y) = N(0,1)parameters x= y+ expansion p(z)=N(x, )noise addition x z l x y

5
p(z|x,l)=N(T l x, l + ) cy x z l cc c=1c=2c=3 Results: Image motion analysis by PGM

6
p(z|x,l)=N(T l x, l + ) cy x z l cc c=1c=2c=3 Results: Image motion analysis by PGM Video summary Image segmentation Sensor noise removal Image stabilisation

Similar presentations

Presentation is loading. Please wait....

OK

Variational Inference and Variational Message Passing

Variational Inference and Variational Message Passing

© 2018 SlidePlayer.com Inc.

All rights reserved.

Ads by Google

Ppt on review of literature definition Ppt on natural numbers list Ppt on mars exploration rover Ppt on condition based maintenance dod Ppt on importance of sports and games in students life Ppt on carbon and its compounds ppt What does appt only meaning Ppt on coalition government australia Ppt on low power tv transmitter Ppt on diode as rectifier regulator