Undirected Probabilistic Graphical Models (Markov Nets) (Slides from Sam Roweis Lecture)

Slides:



Advertisements
Similar presentations
Markov Networks Alan Ritter.
Advertisements

Discriminative Training of Markov Logic Networks
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Discriminative Structure and Parameter.
Online Max-Margin Weight Learning for Markov Logic Networks Tuyen N. Huynh and Raymond J. Mooney Machine Learning Group Department of Computer Science.
CPSC 322, Lecture 30Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 30 March, 25, 2015 Slide source: from Pedro Domingos UW.
Exact Inference in Bayes Nets
Markov Logic Networks: Exploring their Application to Social Network Analysis Parag Singla Dept. of Computer Science and Engineering Indian Institute of.
Supervised Learning Recap
Undirected Probabilistic Graphical Models (Markov Nets) (Slides from Sam Roweis)
Markov Logic: Combining Logic and Probability Parag Singla Dept. of Computer Science & Engineering Indian Institute of Technology Delhi.
Review Markov Logic Networks Mathew Richardson Pedro Domingos Xinran(Sean) Luo, u
Practical Statistical Relational AI Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
Markov Networks.
Unifying Logical and Statistical AI Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint work with Jesse Davis, Stanley Kok,
Markov Logic Networks Hao Wu Mariyam Khalid. Motivation.
Speaker:Benedict Fehringer Seminar:Probabilistic Models for Information Extraction by Dr. Martin Theobald and Maximilian Dylla Based on Richards, M., and.
SAT ∩ AI Henry Kautz University of Rochester. Outline Ancient History: Planning as Satisfiability The Future: Markov Logic.
Chapter 8-3 Markov Random Fields 1. Topics 1. Introduction 1. Undirected Graphical Models 2. Terminology 2. Conditional Independence 3. Factorization.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
11/16: After Sanity Test  Post-mortem  Project presentations in the last 2-3 classes  Start of Statistical Learning.
Statistical Relational Learning Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
Inference. Overview The MC-SAT algorithm Knowledge-based model construction Lazy inference Lifted inference.
Unifying Logical and Statistical AI Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint work with Stanley Kok, Daniel Lowd,
Relational Models. CSE 515 in One Slide We will learn to: Put probability distributions on everything Learn them from data Do inference with them.
Markov Logic Networks: A Unified Approach To Language Processing Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint work with.
Markov Logic: A Simple and Powerful Unification Of Logic and Probability Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint.
Learning, Logic, and Probability: A Unified View Pedro Domingos Dept. Computer Science & Eng. University of Washington (Joint work with Stanley Kok, Matt.
. Approximate Inference Slides by Nir Friedman. When can we hope to approximate? Two situations: u Highly stochastic distributions “Far” evidence is discarded.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1 Learning the Structure of Markov Logic Networks Stanley Kok & Pedro Domingos Dept. of Computer Science and Eng. University of Washington.
Statistical Relational Learning Pedro Domingos Dept. Computer Science & Eng. University of Washington.
Markov Logic Parag Singla Dept. of Computer Science University of Texas, Austin.
Markov Logic: A Unifying Language for Information and Knowledge Management Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint.
1 Naïve Bayes Models for Probability Estimation Daniel Lowd University of Washington (Joint work with Pedro Domingos)
Machine Learning For the Web: A Unified View Pedro Domingos Dept. of Computer Science & Eng. University of Washington Includes joint work with Stanley.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Markov Logic And other SRL Approaches
Statistical Modeling Of Relational Data Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
Markov Random Fields Probabilistic Models for Images
Markov Logic Networks Pedro Domingos Dept. Computer Science & Eng. University of Washington (Joint work with Matt Richardson)
First-Order Logic and Inductive Logic Programming.
1 Markov Logic Stanley Kok Dept. of Computer Science & Eng. University of Washington Joint work with Pedro Domingos, Daniel Lowd, Hoifung Poon, Matt Richardson,
CPSC 322, Lecture 31Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 33 Nov, 25, 2015 Slide source: from Pedro Domingos UW & Markov.
Lecture 2: Statistical learning primer for biologists
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
CPSC 322, Lecture 30Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 30 Nov, 23, 2015 Slide source: from Pedro Domingos UW.
Markov Logic Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
Happy Mittal (Joint work with Prasoon Goyal, Parag Singla and Vibhav Gogate) IIT Delhi New Rules for Domain Independent Lifted.
Markov Logic: A Representation Language for Natural Language Semantics Pedro Domingos Dept. Computer Science & Eng. University of Washington (Based on.
Progress Report ekker. Problem Definition In cases such as object recognition, we can not include all possible objects for training. So transfer learning.
First Order Representations and Learning coming up later: scalability!
CSC Lecture 23: Sigmoid Belief Nets and the wake-sleep algorithm Geoffrey Hinton.
Scalable Statistical Relational Learning for NLP William Y. Wang William W. Cohen Machine Learning Dept and Language Technologies Inst. joint work with:
Probabilistic Reasoning Inference and Relational Bayesian Networks.
New Rules for Domain Independent Lifted MAP Inference
An Introduction to Markov Logic Networks in Knowledge Bases
Inference in Bayesian Networks
Markov Logic Networks for NLP CSCI-GA.2591
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 30
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 29
Logic for Artificial Intelligence
Markov Networks.
Lifted First-Order Probabilistic Inference [de Salvo Braz, Amir, and Roth, 2005] Daniel Lowd 5/11/2005.
Learning Markov Networks
Markov Networks.
Unifying Variational and GBP Learning Parameters of MNs EM for BNs
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Mostly pilfered from Pedro’s slides
Markov Networks.
Presentation transcript:

Undirected Probabilistic Graphical Models (Markov Nets) (Slides from Sam Roweis Lecture)

Connection to MCMC:  MCMC requires sampling a node given its markov blanket  Need to use P(x|MB(x)). For Bayes nets MB(x) contains more nodes than are mentioned in the local distribution CPT(x)  For Markov nets,

 Because neighbor relation is symmetric nodes xi and xj are both neighbors of each other.. In contrast, note that in Bayes Nets, CPTs can be filled with any real numbers between 0 and 1, and we can be sure the ensuing product will define a valid joint distribution!

12/2  All project presentations on 12/14 (10min each)  All project reports due on 12/14  On 12/7, we will read and discuss MLN paper Today: Complete discussion of Markov Nets; Start towards MLN

A B C D Qn: What is the most likely configuration of A&B? Factor says a=b=0 But, marginal says a=0;b=1! Moral: Factors are not marginals! Although A,B would Like to agree, B&C Need to agree, C&D need to disagree And D&A need to agree.and the latter three have Higher weights! Mr. & Mrs. Smith example Okay, you convinced me that given any potentials we will have a consistent Joint. But given any joint, will there be a potentials I can provide?  Hammersley-Clifford theorem… We can have potentials on any cliques—not just the maximal ones. So, for example we can have a potential on A in addition to the other four pairwise potentials

Markov Networks Undirected graphical models Cancer CoughAsthma Smoking Potential functions defined over cliques SmokingCancer Ф(S,C) False 4.5 FalseTrue 4.5 TrueFalse 2.7 True 4.5

Log-Linear models for Markov Nets A B C D Factors are “functions” over their domains Log linear model consists of  Features f i (D i ) (functions over domains)  Weights w i for features s.t. Without loss of generality!

Markov Networks Undirected graphical models Log-linear model: Weight of Feature iFeature i Cancer CoughAsthma Smoking

Markov Nets vs. Bayes Nets PropertyMarkov NetsBayes Nets FormProd. potentials PotentialsArbitraryCond. probabilities CyclesAllowedForbidden Partition func.Z = ? globalZ = 1 local Indep. checkGraph separationD-separation Indep. props.Some InferenceMCMC, BP, etc.Convert to Markov

Inference in Markov Networks Goal: Compute marginals & conditionals of Exact inference is #P-complete Most BN inference approaches work for MNs too – Variable Elimination used factor multiplication—and should work without change.. Conditioning on Markov blanket is easy: Gibbs sampling exploits this

MCMC: Gibbs Sampling state ← random truth assignment for i ← 1 to num-samples do for each variable x sample x according to P(x|neighbors(x)) state ← state with new value of x P(F) ← fraction of states in which F is true

Other Inference Methods Many variations of MCMC Belief propagation (sum-product) Variational approximation Exact methods

Overview Motivation Foundational areas – Probabilistic inference – Statistical learning – Logical inference – Inductive logic programming Putting the pieces together Applications

Learning Markov Networks Learning parameters (weights) – Generatively – Discriminatively Learning structure (features) Easy Case: Assume complete data (If not: EM versions of algorithms)

Entanglement in log likelihood… abc

Learning for log-linear formulation Use gradient ascent Unimodal, because Hessian is Co-variance matrix over features What is the expected Value of the feature given the current parameterization of the network? Requires inference to answer (inference at every iteration— sort of like EM  )

Why should we spend so much time computing gradient? Given that gradient is being used only in doing the gradient ascent iteration, it might look as if we should just be able to approximate it in any which way – Afterall, we are going to take a step with some arbitrary step size anyway....But the thing to keep in mind is that the gradient is a vector. We are talking not just of magnitude but direction. A mistake in magnitude can change the direction of the vector and push the search into a completely wrong direction…

Generative Weight Learning Maximize likelihood or posterior probability Numerical optimization (gradient or 2 nd order) No local maxima Requires inference at each step (slow!) No. of times feature i is true in data Expected no. times feature i is true according to model

Alternative Objectives to maximize.. Since log-likelihood requires network inference to compute the derivative, we might want to focus on other objectives whose gradients are easier to compute (and which also – hopefully—have optima at the same parameter values). Two options: – Pseudo Likelihood – Contrastive Divergence Given a single data instance  log-likelihood is Log prob of data Log prob of all other possible data instances (w.r.t. current  Maximize the distance (“increase the divergence”) Pick a sample of typical other instances (need to sample from P  Run MCMC initializing with the data..) Compute likelihood of each possible data instance just using markov blanket (approximate chain rule)

Pseudo-Likelihood Likelihood of each variable given its neighbors in the data Does not require inference at each step Consistent estimator Widely used in vision, spatial statistics, etc. But PL parameters may not work well for long inference chains [Which can lead to disasterous results]

Discriminative Weight Learning Maximize conditional likelihood of query ( y ) given evidence ( x ) Approximate expected counts by counts in MAP state of y given x No. of true groundings of clause i in data Expected no. true groundings according to model

Structure Learning How to learn the structure of a Markov network? – … not too different from learning structure for a Bayes network: discrete search through space of possible graphs, trying to maximize data probability….

MLNs: Points to ponder Compared to ground representations, MLNs have easier learning but equal harder inference – MLNs need to learn significantly fewer parameters than a ground network of similar size – MLNs may be compelled to exploit the “relational” structure and thus may spend time inventing lifted inference methods Inference approaches Learning – Parameter Why Pseudo Likelihood? – Structure—implies learning clauses.. (what ILP does) Connection to Dynamic Bayes Nets? Relational

Markov Logic: Intuition A logical KB is a set of hard constraints on the set of possible worlds Let’s make them soft constraints: When a world violates a formula, It becomes less probable, not impossible Give each formula a weight (Higher weight  Stronger constraint)

Markov Logic: Definition A Markov Logic Network (MLN) is a set of pairs (F, w) where – F is a formula in first-order logic – w is a real number Together with a set of constants, it defines a Markov network with – One node for each grounding of each predicate in the MLN – One feature for each grounding of each formula F in the MLN, with the corresponding weight w

Example: Friends & Smokers

Two constants: Anna (A) and Bob (B)

Example: Friends & Smokers Cancer(A) Smokes(A)Smokes(B) Cancer(B) Two constants: Anna (A) and Bob (B)

Example: Friends & Smokers Cancer(A) Smokes(A)Friends(A,A) Friends(B,A) Smokes(B) Friends(A,B) Cancer(B) Friends(B,B) Two constants: Anna (A) and Bob (B)

Example: Friends & Smokers Cancer(A) Smokes(A)Friends(A,A) Friends(B,A) Smokes(B) Friends(A,B) Cancer(B) Friends(B,B) Two constants: Anna (A) and Bob (B)

Example: Friends & Smokers Cancer(A) Smokes(A)Friends(A,A) Friends(B,A) Smokes(B) Friends(A,B) Cancer(B) Friends(B,B) Two constants: Anna (A) and Bob (B)

Markov Logic Networks MLN is template for ground Markov nets Probability of a world x : Typed variables and constants greatly reduce size of ground Markov net Functions, existential quantifiers, etc. Infinite and continuous domains Weight of formula iNo. of true groundings of formula i in x

Relation to Statistical Models Special cases: – Markov networks – Markov random fields – Bayesian networks – Log-linear models – Exponential models – Max. entropy models – Gibbs distributions – Boltzmann machines – Logistic regression – Hidden Markov models – Conditional random fields Obtained by making all predicates zero-arity Markov logic allows objects to be interdependent (non-i.i.d.)

Relation to First-Order Logic Infinite weights  First-order logic Satisfiable KB, positive weights  Satisfying assignments = Modes of distribution Markov logic allows contradictions between formulas

MAP/MPE Inference Problem: Find most likely state of world given evidence QueryEvidence

MAP/MPE Inference Problem: Find most likely state of world given evidence

MAP/MPE Inference Problem: Find most likely state of world given evidence

MAP/MPE Inference Problem: Find most likely state of world given evidence This is just the weighted MaxSAT problem Use weighted SAT solver (e.g., MaxWalkSAT [Kautz et al., 1997] ) Potentially faster than logical inference (!)

The MaxWalkSAT Algorithm for i ← 1 to max-tries do solution = random truth assignment for j ← 1 to max-flips do if ∑ weights(sat. clauses) > threshold then return solution c ← random unsatisfied clause with probability p flip a random variable in c else flip variable in c that maximizes ∑ weights(sat. clauses) return failure, best solution found

But … Memory Explosion Problem: If there are n constants and the highest clause arity is c, the ground network requires O(n ) memory Solution: Exploit sparseness; ground clauses lazily → LazySAT algorithm [Singla & Domingos, 2006] c

Computing Probabilities P(Formula|MLN,C) = ? MCMC: Sample worlds, check formula holds P(Formula1|Formula2,MLN,C) = ? If Formula2 = Conjunction of ground atoms – First construct min subset of network necessary to answer query (generalization of KBMC) – Then apply MCMC (or other) Can also do lifted inference [Braz et al, 2005]

Ground Network Construction network ← Ø queue ← query nodes repeat node ← front(queue) remove node from queue add node to network if node not in evidence then add neighbors(node) to queue until queue = Ø

But … Insufficient for Logic Problem: Deterministic dependencies break MCMC Near-deterministic ones make it very slow Solution: Combine MCMC and WalkSAT → MC-SAT algorithm [Poon & Domingos, 2006]

Learning Data is a relational database Closed world assumption (if not: EM) Learning parameters (weights) Learning structure (formulas)

Parameter tying: Groundings of same clause Generative learning: Pseudo-likelihood Discriminative learning: Cond. likelihood, use MC-SAT or MaxWalkSAT for inference Weight Learning No. of times clause i is true in data Expected no. times clause i is true according to MLN

Structure Learning Generalizes feature induction in Markov nets Any inductive logic programming approach can be used, but... Goal is to induce any clauses, not just Horn Evaluation function should be likelihood Requires learning weights for each candidate Turns out not to be bottleneck Bottleneck is counting clause groundings Solution: Subsampling

Structure Learning Initial state: Unit clauses or hand-coded KB Operators: Add/remove literal, flip sign Evaluation function: Pseudo-likelihood + Structure prior Search: Beam, shortest-first, bottom-up [Kok & Domingos, 2005; Mihalkova & Mooney, 2007]

Alchemy Open-source software including: Full first-order logic syntax Generative & discriminative weight learning Structure learning Weighted satisfiability and MCMC Programming language features alchemy.cs.washington.edu

AlchemyPrologBUGS Represent- ation F.O. Logic + Markov nets Horn clauses Bayes nets InferenceModel check- ing, MC-SAT Theorem proving Gibbs sampling LearningParameters & structure NoParams. UncertaintyYesNoYes RelationalYes No