第十讲 概率图模型导论 Chapter 10 Introduction to Probabilistic Graphical Models

Slides:



Advertisements
Similar presentations
Introduction to Graphical Models Brookes Vision Lab Reading Group.
Advertisements

CS188: Computational Models of Human Behavior
Autonomic Scaling of Cloud Computing Resources
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
Markov Networks Alan Ritter.
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Bayesian Network and Influence Diagram A Guide to Construction And Analysis.
Graphical Models BRML Chapter 4 1. the zoo of graphical models Markov networks Belief networks Chain graphs (Belief and Markov ) Factor graphs =>they.
Conditional Random Fields - A probabilistic graphical model Stefan Mutter Machine Learning Group Conditional Random Fields - A probabilistic graphical.
Introduction of Probabilistic Reasoning and Bayesian Networks
Consistent probabilistic outputs for protein function prediction William Stafford Noble Department of Genome Sciences Department of Computer Science and.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Introduction to probability theory and graphical models Translational Neuroimaging Seminar on Bayesian Inference Spring 2013 Jakob Heinzle Translational.
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Today Logistic Regression Decision Trees Redux Graphical Models
Bayesian Networks Alan Ritter.
CPSC 422, Lecture 18Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Feb, 25, 2015 Slide Sources Raymond J. Mooney University of.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Causal Modeling for Anomaly Detection Andrew Arnold Machine Learning Department, Carnegie Mellon University Summer Project with Naoki Abe Predictive Modeling.
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Bayes’ Nets  A Bayes’ net is an efficient encoding of a probabilistic model of a domain  Questions we can ask:  Inference: given a fixed BN, what is.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
A Brief Introduction to Graphical Models
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
Probabilistic graphical models. Graphical models are a marriage between probability theory and graph theory (Michael Jordan, 1998) A compact representation.
A Markov Random Field Model for Term Dependencies Donald Metzler W. Bruce Croft Present by Chia-Hao Lee.
Perceptual and Sensory Augmented Computing Machine Learning, Summer’11 Machine Learning – Lecture 13 Introduction to Graphical Models Bastian.
Bayesian Network By DengKe Dong. Key Points Today  Intro to Graphical Model  Conditional Independence  Intro to Bayesian Network  Reasoning BN: D-Separation.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Introduction to Bayesian Networks
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 28 of 41 Friday, 22 October.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Graphical Models over Multiple Strings Markus Dreyer and Jason Eisner Dept. of Computer Science, Johns Hopkins University EMNLP 2009 Presented by Ji Zongcheng.
Announcements Project 4: Ghostbusters Homework 7
LAC group, 16/06/2011. So far...  Directed graphical models  Bayesian Networks Useful because both the structure and the parameters provide a natural.
Presented by Jian-Shiun Tzeng 5/7/2009 Conditional Random Fields: An Introduction Hanna M. Wallach University of Pennsylvania CIS Technical Report MS-CIS
Slides for “Data Mining” by I. H. Witten and E. Frank.
An Introduction to Variational Methods for Graphical Models
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
Probabilistic Graphical Models seminar 15/16 ( ) Haim Kaplan Tel Aviv University.
Lecture 2: Statistical learning primer for biologists
Indexing Correlated Probabilistic Databases Bhargav Kanagal, Amol Deshpande University of Maryland, College Park, USA SIGMOD Presented.
CPSC 422, Lecture 17Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 17 Oct, 19, 2015 Slide Sources D. Koller, Stanford CS - Probabilistic.
Maximum Entropy Model, Bayesian Networks, HMM, Markov Random Fields, (Hidden/Segmental) Conditional Random Fields.
Christopher M. Bishop, Pattern Recognition and Machine Learning 1.
Machine Learning – Lecture 11
Pattern Recognition and Machine Learning
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
04/21/2005 CS673 1 Being Bayesian About Network Structure A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller.
Introduction on Graphic Models
Markov Random Fields in Vision
An Algorithm to Learn the Structure of a Bayesian Network Çiğdem Gündüz Olcay Taner Yıldız Ethem Alpaydın Computer Engineering Taner Bilgiç Industrial.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
Artificial Intelligence Bayes’ Nets: Independence Instructors: David Suter and Qince Li Course Harbin Institute of Technology [Many slides.
Integrative Genomics I BME 230. Probabilistic Networks Incorporate uncertainty explicitly Capture sparseness of wiring Incorporate multiple kinds of data.
INTRODUCTION TO Machine Learning 2nd Edition
Bayesian Framework Finding the best model Minimizing model complexity
CHAPTER 16: Graphical Models
Maximum Expected Utility
ICS 280 Learning in Graphical Models
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 17
CAP 5636 – Advanced Artificial Intelligence
Graphical Models.
Markov Random Fields Presented by: Vladan Radosavljevic.
CS 188: Artificial Intelligence
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
Chapter 14 February 26, 2004.
Presentation transcript:

第十讲 概率图模型导论 Chapter 10 Introduction to Probabilistic Graphical Models 浙江大学计算机学院《人工智能引论》课件 第十讲 概率图模型导论 Chapter 10 Introduction to Probabilistic Graphical Models Weike Pan, and Congfu Xu {panweike, xucongfu}@zju.edu.cn Institute of Artificial Intelligence College of Computer Science, Zhejiang University October 12, 2006

References An Introduction to Probabilistic Graphical Models. Michael I. Jordan. http://www.cs.berkeley.edu/~jordan/graphical.html

Outline Preparations Probabilistic Graphical Models (PGM) Directed PGM Undirected PGM Insights of PGM

Outline Preparations Probabilistic Graphical Models (PGM) PGM “is” a universal model Different thoughts of machine learning Different training approaches Different data types Bayesian Framework Chain rules of probability theory Conditional Independence Probabilistic Graphical Models (PGM) Directed PGM Undirected PGM Insights of PGM

Different thoughts of machine learning Statistics (modeling uncertainty, detailed information) vs. Logics (modeling complexity, high level information) Unifying Logical and Statistical AI. Pedro Domingos, University of Washington. AAAI 2006. Speech: Statistical information (Acoustic model + Language model + Affect model…) + High level information (Expert/Logics) Machine Learning: supervised, unsupervised, reinforcement learning, multi-instance learning Thoughts: Statistics, Logics Trends: unifying the two thoughts together

Different training approaches Maximum Likelihood Training: MAP (Maximum a Posteriori) vs. Discriminative Training: Maximum Margin (SVM) Speech: classical combination – Maximum Likelihood + Discriminative Training

Different data types Directed acyclic graph (Bayesian Networks, BN) Modeling asymmetric effects and dependencies: causal/temporal dependence (e.g. speech analysis, DNA sequence analysis…) Undirected graph (Markov Random Fields, MRF) Modeling symmetric effects and dependencies: spatial dependence (e.g. image analysis…)

PGM “is” a universal model To model both temporal and spatial data, by unifying Thoughts: Statistics + Logics Approaches: Maximum Likelihood Training + Discriminative Training Further more, the directed and undirected models together provide modeling power beyond that which could be provided by either alone.

Bayesian Framework Problem description Bayesian rule Observation  Conclusion (classification or prediction) Bayesian rule Likelihood Priori probability Observation A posteriori probability Class i Normalization factor What we care is the conditional probability, and it’s is a ratio of two marginal probabilities.

Chain rules of probability theory

Conditional Independence

Outline Preparations Probabilistic Graphical Models (PGM) Directed PGM Undirected PGM Insights of PGM

PGM Nodes represent random variables/states The missing arcs represent conditional independence assumptions The graph structure implies the decomposition

Directed PGM (BN) Probability Distribution Queries Representation Implementation Conditional Independence Interpretation

Probability Distribution Definition of Joint Probability Distribution Check:

Representation Graphical models represent joint probability distributions more economically, using a set of “local” relationships among variables.

Conditional Independence (basic) Interpret missing edges in terms of conditional independence Assert the conditional independence of a node from its ancestors, conditional on its parents.

Conditional Independence (3 canonical graphs) Marginal Independence Classical Markov chain “Past”, “present”, “future” Common cause Y “explains” all the dependencies between X and Z Common effect Multiple, competing explanation

Conditional Independence (check) Bayes ball algorithm (rules) One incoming arrow and one outgoing arrow Two outgoing arrows Two incoming arrows Check through reachability

Outline Preparations Probabilistic Graphical Models (PGM) Directed PGM Undirected PGM Insights of PGM

Undirected PGM (MRF) Probability Distribution Queries Representation Implementation Conditional Independence Interpretation

Probability Distribution(1) Clique A clique of a graph is a fully-connected subset of nodes. Local functions should not be defined on domains of nodes that extend beyond the boundaries of cliques. Maximal cliques The maximal cliques of a graph are the cliques that cannot be extended to include additional nodes without losing the probability of being fully connected. We restrict ourselves to maximal cliques without loss of generality, as it captures all possible dependencies. Potential function (local parameterization) : potential function on the possible realizations of the maximal clique

Probability Distribution(2) Maximal cliques

Probability Distribution(3) Joint probability distribution Normalization factor Boltzman distribution

Conditional Independence It’s a “reachability” problem in graph theory.

Representation

Outline Preparations Probabilistic Graphical Models (PGM) Directed PGM Undirected PGM Insights of PGM

Insights of PGM (Michael I. Jordan) Probabilistic Graphical Models are a marriage between probability theory and graph theory. A graphical model can be thought of as a probabilistic database, a machine that can answer “queries” regarding the values of sets of random variables. We build up the database in pieces, using probability theory to ensure that the pieces have a consistent overall interpretation. Probability theory also justifies the inferential machinery that allows the pieces to be put together “on the fly” to answer the queries. In principle, all “queries” of a probabilistic database can be answered if we have in hand the joint probability distribution.

Insights of PGM (data structure & algorithm) A graphical model is a natural/perfect tool for representation(数据结构) and inference (算法).

Thanks!