Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.

Slides:



Advertisements
Similar presentations
A Tutorial on Learning with Bayesian Networks
Advertisements

Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Review of Probability. Definitions (1) Quiz 1.Let’s say I have a random variable X for a coin, with event space {H, T}. If the probability P(X=H) is.
Bayesian Estimation in MARK
Dynamic Bayesian Networks (DBNs)
Gibbs Sampling Qianji Zheng Oct. 5th, 2010.
Introduction of Probabilistic Reasoning and Bayesian Networks
CHAPTER 16 MARKOV CHAIN MONTE CARLO
BAYESIAN INFERENCE Sampling techniques
Exact Inference (Last Class) variable elimination  polytrees (directed graph with at most one undirected path between any two vertices; subset of DAGs)
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
End of Chapter 8 Neil Weisenfeld March 28, 2005.
CS 188: Artificial Intelligence Spring 2007 Lecture 14: Bayes Nets III 3/1/2007 Srini Narayanan – ICSI and UC Berkeley.
CS 188: Artificial Intelligence Fall 2006 Lecture 17: Bayes Nets III 10/26/2006 Dan Klein – UC Berkeley.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Today Logistic Regression Decision Trees Redux Graphical Models
Announcements Homework 8 is out Final Contest (Optional)
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Artificial Intelligence CS 165A Tuesday, November 27, 2007  Probabilistic Reasoning (Ch 14)
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Image Analysis and Markov Random Fields (MRFs) Quanren Xiong.
Bayesian networks Chapter 14. Outline Syntax Semantics.
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Randomized Algorithms for Bayesian Hierarchical Clustering
Bayes’ Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available.
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Exact Inference (Last Class) Variable elimination  polytrees (directed graph with at most one undirected path between any two vertices; subset of DAGs)
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
CHAPTER 5 Probability Theory (continued) Introduction to Bayesian Networks.
Lecture 2: Statistical learning primer for biologists
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
CS 188: Artificial Intelligence Bayes Nets: Approximate Inference Instructor: Stuart Russell--- University of California, Berkeley.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Quick Warm-Up Suppose we have a biased coin that comes up heads with some unknown probability p; how can we use it to produce random bits with probabilities.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
CS774. Markov Random Field : Theory and Application Lecture 15 Kyomin Jung KAIST Oct
Pattern Recognition and Machine Learning
CPSC 7373: Artificial Intelligence Lecture 5: Probabilistic Inference Jiang Bian, Fall 2012 University of Arkansas at Little Rock.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
CS 416 Artificial Intelligence Lecture 15 Uncertainty Chapter 14 Lecture 15 Uncertainty Chapter 14.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Daphne Koller Overview Conditional Probability Queries Probabilistic Graphical Models Inference.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
CS 2750: Machine Learning Directed Graphical Models
Gibbs sampling.
Artificial Intelligence
CS 4/527: Artificial Intelligence
Prof. Adriana Kovashka University of Pittsburgh April 4, 2017
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
CAP 5636 – Advanced Artificial Intelligence
Markov Networks.
CSE-490DF Robotics Capstone
Instructors: Fei Fang (This Lecture) and Dave Touretzky
Ch13 Empirical Methods.
CS 188: Artificial Intelligence
Class #19 – Tuesday, November 3
CS 188: Artificial Intelligence Fall 2008
Class #16 – Tuesday, October 26
Approximate Inference by Sampling
Markov Networks.
Presentation transcript:

Probabilistic models Jouni Tuomisto THL

Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets

Deterministic models with probabilistic parameters Inputs are uncertain, but causal relations are assumed certain. Works well with established situations, especially if physical foundations. Exposure = ∑ (c i t i ) / ∑ t i –i = microenvironment –c = concentration –t = time

Functional vs. probabilistic dependency Va1=2.54*Ch1^2 Va2=normal(2.54*Ch1^2,2)

Hierarchical Bayesian models Relations are probabilistic Gibbs Sampler Another MCMC (Markov chain Monte Carlo) Method Update a single parameter at a time Sample from conditional distribution When other parameters are fixed

Gibbs sampling To introduce the Gibbs sampler, consider a bivariate random variable (x; y), and suppose we wish to compute one or both marginals, p(x) and p(y). The idea behind the sampler is that it is far easier to consider a sequence of conditional distributions, p(x | y) and p(y | x), than it is to obtain the marginal by integration of the joint density p(x; y), e.g., –p(x) = ∫ p(x; y)dy.

Gibbs sampling in practice The sampler starts with some initial value y0 for y and obtains x0 by generating a random variable from the conditional distribution p(x | y = y0). The sampler then uses x0 to generate a new value of y1, drawing from the conditional distribution based on the value x0, p(y j x = x0). The sampler proceeds as follows xi ≈ p(x | y = yi-1) (proportionality) yi ≈ p(y | x = xi) Repeating this process k times, generates a Gibbs sequence of length k, where a subset of points (xj; yj) for 1 ≤ j ≤ m < k are taken as our simulated draws from the full joint distribution.

Hierarchical model with parameters and hyperparameters A useful graphical tool for representing hierarchical Bayes models is the directed acyclic graph, or DAG. In this diagram, the likelihood function is represented as the root of the graph; each prior is represented as a separate node pointing to the node that depends on it.directed acyclic graph

Bayesian belief nets Relations are described either with conditional probabilities P(x|y), P(y) or marginal probabilities P(x), P(y) and a rank correlation between them. You need to get the conditional probabilities from somewhere. –Unlike hierarchical Bayes model, belief nets are not developed for updating when new data comes out. The model is used to make inference.

Bayesian belief nets P(sprinkler | rain) P(grass wet | sprinkler, rain) P(rain)