Markov Logic: Combining Logic and Probability Parag Singla Dept. of Computer Science & Engineering Indian Institute of Technology Delhi.

Slides:



Advertisements
Similar presentations
Markov Networks Alan Ritter.
Advertisements

Discriminative Training of Markov Logic Networks
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Discriminative Structure and Parameter.
Bayesian Abductive Logic Programs Sindhu Raghavan Raymond J. Mooney The University of Texas at Austin 1.
Online Max-Margin Weight Learning for Markov Logic Networks Tuyen N. Huynh and Raymond J. Mooney Machine Learning Group Department of Computer Science.
CPSC 322, Lecture 30Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 30 March, 25, 2015 Slide source: from Pedro Domingos UW.
Markov Logic Networks: Exploring their Application to Social Network Analysis Parag Singla Dept. of Computer Science and Engineering Indian Institute of.
Undirected Probabilistic Graphical Models (Markov Nets) (Slides from Sam Roweis Lecture)
Markov Logic Networks Instructor: Pedro Domingos.
Department of Computer Science The University of Texas at Austin Probabilistic Abduction using Markov Logic Networks Rohit J. Kate Raymond J. Mooney.
Undirected Probabilistic Graphical Models (Markov Nets) (Slides from Sam Roweis)
Review Markov Logic Networks Mathew Richardson Pedro Domingos Xinran(Sean) Luo, u
Speeding Up Inference in Markov Logic Networks by Preprocessing to Reduce the Size of the Resulting Grounded Network Jude Shavlik Sriraam Natarajan Computer.
Adbuctive Markov Logic for Plan Recognition Parag Singla & Raymond J. Mooney Dept. of Computer Science University of Texas, Austin.
Markov Networks.
Unifying Logical and Statistical AI Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint work with Jesse Davis, Stanley Kok,
Markov Logic: A Unifying Framework for Statistical Relational Learning Pedro Domingos Matthew Richardson
Markov Logic Networks Hao Wu Mariyam Khalid. Motivation.
Speaker:Benedict Fehringer Seminar:Probabilistic Models for Information Extraction by Dr. Martin Theobald and Maximilian Dylla Based on Richards, M., and.
School of Computing Science Simon Fraser University Vancouver, Canada.
SAT ∩ AI Henry Kautz University of Rochester. Outline Ancient History: Planning as Satisfiability The Future: Markov Logic.
CSE 574 – Artificial Intelligence II Statistical Relational Learning Instructor: Pedro Domingos.
Statistical Relational Learning Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
Recursive Random Fields Daniel Lowd University of Washington June 29th, 2006 (Joint work with Pedro Domingos)
CSE 574: Artificial Intelligence II Statistical Relational Learning Instructor: Pedro Domingos.
Inference. Overview The MC-SAT algorithm Knowledge-based model construction Lazy inference Lifted inference.
Recursive Random Fields Daniel Lowd University of Washington (Joint work with Pedro Domingos)
Unifying Logical and Statistical AI Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint work with Stanley Kok, Daniel Lowd,
Relational Models. CSE 515 in One Slide We will learn to: Put probability distributions on everything Learn them from data Do inference with them.
Markov Logic Networks: A Unified Approach To Language Processing Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint work with.
Markov Logic: A Simple and Powerful Unification Of Logic and Probability Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint.
1 Human Detection under Partial Occlusions using Markov Logic Networks Raghuraman Gopalan and William Schwartz Center for Automation Research University.
Learning, Logic, and Probability: A Unified View Pedro Domingos Dept. Computer Science & Eng. University of Washington (Joint work with Stanley Kok, Matt.
On the Proper Treatment of Quantifiers in Probabilistic Logic Semantics Islam Beltagy and Katrin Erk The University of Texas at Austin IWCS 2015.
CPSC 422, Lecture 18Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Feb, 25, 2015 Slide Sources Raymond J. Mooney University of.
Statistical Relational Learning Pedro Domingos Dept. Computer Science & Eng. University of Washington.
Pedro Domingos Dept. of Computer Science & Eng.
Markov Logic Parag Singla Dept. of Computer Science University of Texas, Austin.
Markov Logic: A Unifying Language for Information and Knowledge Management Pedro Domingos Dept. of Computer Science & Eng. University of Washington Joint.
Machine Learning For the Web: A Unified View Pedro Domingos Dept. of Computer Science & Eng. University of Washington Includes joint work with Stanley.
Markov Logic And other SRL Approaches
Markov Random Fields Probabilistic Models for Images
Markov Logic Networks Pedro Domingos Dept. Computer Science & Eng. University of Washington (Joint work with Matt Richardson)
Modeling Speech Acts and Joint Intentions in Modal Markov Logic Henry Kautz University of Washington.
1 Markov Logic Stanley Kok Dept. of Computer Science & Eng. University of Washington Joint work with Pedro Domingos, Daniel Lowd, Hoifung Poon, Matt Richardson,
CPSC 322, Lecture 31Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 33 Nov, 25, 2015 Slide source: from Pedro Domingos UW & Markov.
Abductive Plan Recognition By Extending Bayesian Logic Programs Sindhu V. Raghavan & Raymond J. Mooney The University of Texas at Austin 1.
CPSC 322, Lecture 30Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 30 Nov, 23, 2015 Slide source: from Pedro Domingos UW.
CPSC 422, Lecture 17Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 17 Oct, 19, 2015 Slide Sources D. Koller, Stanford CS - Probabilistic.
Markov Logic Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
Happy Mittal (Joint work with Prasoon Goyal, Parag Singla and Vibhav Gogate) IIT Delhi New Rules for Domain Independent Lifted.
Markov Logic: A Representation Language for Natural Language Semantics Pedro Domingos Dept. Computer Science & Eng. University of Washington (Based on.
1 Relational Factor Graphs Lin Liao Joint work with Dieter Fox.
Progress Report ekker. Problem Definition In cases such as object recognition, we can not include all possible objects for training. So transfer learning.
First Order Representations and Learning coming up later: scalability!
Scalable Statistical Relational Learning for NLP William Y. Wang William W. Cohen Machine Learning Dept and Language Technologies Inst. joint work with:
Happy Mittal Advisor : Parag Singla IIT Delhi Lifted Inference Rules With Constraints.
New Rules for Domain Independent Lifted MAP Inference
An Introduction to Markov Logic Networks in Knowledge Bases
Scalable Statistical Relational Learning for NLP
Markov Logic Networks for NLP CSCI-GA.2591
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 30
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 29
Discriminative Learning for Markov Logic Networks
Logic for Artificial Intelligence
Lifted First-Order Probabilistic Inference [de Salvo Braz, Amir, and Roth, 2005] Daniel Lowd 5/11/2005.
Expectation-Maximization & Belief Propagation
Mostly pilfered from Pedro’s slides
Markov Networks.
Statistical Relational AI
Presentation transcript:

Markov Logic: Combining Logic and Probability Parag Singla Dept. of Computer Science & Engineering Indian Institute of Technology Delhi

Overview Motivation & Background Markov logic Inference & Learning Abductive Plan Recognition

Social Network and Smoking Behavior SmokingCancer

Social Network and Smoking Behavior Smoking leads toCancer

Social Network and Smoking Behavior Smoking leads toCancer Friendship Similar Smoking Habits

Social Network and Smoking Behavior Smoking leads toCancer Friendship leads to Similar Smoking Habits

Statistical Relational AI Real world problems characterized by Entities and Relationships Uncertain Behavior Relational Models Horn clauses, SQL queries, first-order logic Statistical Models Markov networks, Bayesian networks How to combine the two? Markov Logic Markov Networks + First Order Logic

Statistical Relational AI Probabilistic logic [Nilsson, 1986] Statistics and beliefs [Halpern, 1990] Knowledge-based model construction [Wellman et al., 1992] Stochastic logic programs [Muggleton, 1996] Probabilistic relational models [Friedman et al., 1999] Bayesian Logic Programs [Kersting and De Raedt 2001] Relational Markov networks [Taskar et al., 2002] BLOG [Milch et al., 2005] Markov logic [Richardson & Domingos, 2006]

First-Order Logic Constants, variables, functions, predicates Anil, x, MotherOf(x), Friends(x,y) Grounding: Replace all variables by constants Friends (Anna, Bob) Formula: Predicates connected by operators Smokes(x)  Cancer(x) Knowledge Base (KB): A set of formulas Can be equivalently converted into a clausal form World: Assignment of truth values to all ground atoms

First-Order Logic Deal with finite first-order logic Assumptions Unique Names Domain Closure Known Functions

Markov Networks Undirected graphical models Log-linear model: Weight of Feature iFeature i Cancer CoughAsthma Smoking

Overview Motivation & Background Markov logic Inference & Learning Abductive Plan Recognition

Markov Logic [Richardson & Domingos 06] A logical KB is a set of hard constraints on the set of possible worlds Let’s make them soft constraints: When a world violates a formula, It becomes less probable, not impossible Give each formula a weight (Higher weight  Stronger constraint)

Definition A Markov Logic Network (MLN) is a set of pairs (F, w) where F is a formula in first-order logic w is a real number Together with a finite set of constants, it defines a Markov network with One node for each grounding of each predicate in the MLN One feature for each grounding of each formula F in the MLN, with the corresponding weight w

Example: Friends & Smokers

Two constants: Anil (A) and Bunty (B)

Example: Friends & Smokers Cancer(A) Smokes(A)Smokes(B) Cancer(B) Two constants: Anil (A) and Bunty (B)

Example: Friends & Smokers Cancer(A) Smokes(A)Friends(A,A) Friends(B,A) Smokes(B) Friends(A,B) Cancer(B) Friends(B,B) Two constants: Anil (A) and Bunty (B)

Example: Friends & Smokers Cancer(A) Smokes(A)Friends(A,A) Friends(B,A) Smokes(B) Friends(A,B) Cancer(B) Friends(B,B) Two constants: Anil (A) and Bunty (B)

Example: Friends & Smokers Cancer(A) Smokes(A)Friends(A,A) Friends(B,A) Smokes(B) Friends(A,B) Cancer(B) Friends(B,B) Two constants: Anil (A) and Bunty (B)

Example: Friends & Smokers Cancer(A) Smokes(A)Friends(A,A) Friends(B,A) Smokes(B) Friends(A,B) Cancer(B) Friends(B,B) Two constants: Anil (A) and Bunty (B) State of the World  {0,1} Assignment to the nodes

Markov Logic Networks MLN is template for ground Markov nets Probability of a world x :

Markov Logic Networks MLN is template for ground Markov nets Probability of a world x : Weight of formula iNo. of true groundings of formula i in x

Relation to Statistical Models Special cases: Markov networks Markov random fields Bayesian networks Log-linear models Exponential models Logistic regression Hidden Markov models Conditional random fields Obtained by making all predicates zero-arity

Relation to First-Order Logic Infinite weights  First-order logic Satisfiable KB, positive weights  Satisfying assignments = Modes of distribution Markov logic allows contradictions between formulas Relaxing Assumptions Known Functions (Markov Logic in Infinite Domains) [Singla & Domingos 07] Unique Names (Entity Resolution with Markov Logic) [Singla & Domingos 06]

Overview Motivation & Background Markov logic Inference & Learning Abductive Plan Recognition

Inference Cancer(A) Smokes(A)? Friends(A,A) Friends(B,A) Smokes(B)? Friends(A,B) Friends(B,B) Cancer(B) blue ? – non-evidence (unknown) green/orange – evidence (known)

MPE Inference Problem: Find most likely state of world given evidence QueryEvidence

MPE Inference Problem: Find most likely state of world given evidence

MPE Inference Problem: Find most likely state of world given evidence

MPE Inference Problem: Find most likely state of world given evidence This is just the weighted MaxSAT problem Use weighted SAT solver (e.g., MaxWalkSAT [Kautz et al. 97] ) Lazy Grounding of Clauses: LazySAT [Singla & Domingos 06]

Marginal Inference Problem: Find the probability of query atoms given evidence QueryEvidence

Marginal Inference Problem: Find the probability of query atoms given evidence QueryEvidence Computing Z x takes exponential time!

Marginal Inference Problem: Find the probability of query atoms given evidence QueryEvidence Approximate Inference: Gibbs Sampling, Message Passing [Richardson & Domingos 06, Poon & Domingos 06, Singla & Domingos 08]

Learning Parameters Three constants: Anil, Bunty, Chaya

Learning Parameters Smokes Smokes(Anil) Smokes(Bunty) Closed World Assumption: Anything not in the database is assumed false. Three constants: Anil, Bunty, Chaya Cancer Cancer(Anil) Cancer(Bunty) Friends Friends(Anil, Bunty) Friends(Bunty, Anil) Friends(Anil, Chaya) Friends(Chaya, Anil)

Learning Parameters Smokes Smokes(Anil) Smokes(Bunty) Three constants: Anil, Bunty, Chaya Cancer Cancer(Anil) Cancer(Bunty) Friends Friends(Anil, Bunty) Friends(Bunty, Anil) Friends(Anil, Chaya) Friends(Chaya, Anil) Maximize the Likelihood: Use Gradient Based Approaches [Singla & Domingos 05, Lowd & Domingos 07]

Learning Structure Smokes Smokes(Anil) Smokes(Bunty) Cancer Cancer(Anil) Cancer(Bunty) Friends Friends(Anil, Bob) Friends(Bunty, Anil) Friends(Anil, Chaya) Friends(Chaya, Anil) Three constants: Anil, Bunty, Chaya Can we learn the set of the formulas in the MLN?

Learning Structure Can we refine the set of the formulas in the MLN? Smokes Smokes(Anil) Smokes(Bunty) Cancer Cancer(Anil) Cancer(Bunty) Friends Friends(Anil, Bob) Friends(Bunty, Anil) Friends(Anil, Chaya) Friends(Chaya, Anil) Three constants: Anil, Bunty, Chaya

Learning Structure Can we refine the set of the formulas in the MLN? Smokes Smokes(Anil) Smokes(Bunty) Cancer Cancer(Anil) Cancer(Bunty) Friends Friends(Anil, Bob) Friends(Bunty, Anil) Friends(Anil, Chaya) Friends(Chaya, Anil) Three constants: Anil, Bunty, Chaya

Learning Structure ILP style search for formuals [Kok & Domingos 05, 07, 09, 10] Smokes Smokes(Anil) Smokes(Bunty) Cancer Cancer(Anil) Cancer(Bunty) Friends Friends(Anil, Bob) Friends(Bunty, Anil) Friends(Anil, Chaya) Friends(Chaya, Anil) Three constants: Anil, Bunty, Chaya

Alchemy Open-source software including: Full first-order logic syntax Inference algorithms Parameter & structure learning algorithms alchemy.cs.washington.edu

Overview Motivation & Background Markov logic Inference & Learning Abductive Plan Recognition

Applications Web-mining Collective Classification Link Prediction Information retrieval Entity resolution Activity Recognition Image Segmentation & De-noising Social Network Analysis Computational Biology Natural Language Processing Robot mapping Abductive Plan Recognition More..

Abduction Abduction: Given the observations and the background, find the best explanation Given: Background knowledge (B) A set of observations (O) To Find: A hypothesis, H, a set of assumptions B  H  , B  H  O

Plan Recognition Given planning knowledge and a set of low- level actions, identify the top level plan Involves abductive reasoning B: Planning Knowledge (Background) O: Set of low-level Actions (Observations) H: Top Level Plan (Hypothesis) B  H  , B  H |  O

Plan Recognition Example Emergency Response Domain [Blaylock & Allen 05] Background Knowledge heavy_snow(loc)  drive_hazard(loc)  block_road(loc) accident(loc)  clear_wreck(crew,loc)  block_road(loc) Observation block_road(Plaza) Possible Explanations Heavy Snow? Accident?

Abduction using Markov logic Given heavy_snow(loc)  drive_hazard(loc)  block_road(loc) accdent(loc)  clear_wreck(crew, loc)  block_road(loc) Observation: block_road(plaza)

Abduction using Markov logic Given heavy_snow(loc)  drive_hazard(loc)  block_road(loc) accdent(loc)  clear_wreck(crew, loc)  block_road(loc) Observation: block_road(plaza) Does not work! Rules are true independent of antecedents Need to go from effect to cause

Introducing Hidden Cause heavy_snow(loc)  drive_hazard(loc)  block_road(loc) heavy_snow(loc)  drive_hazard(loc)  rb_C1(loc) rb_C1(loc) Hidden Cause

Introducing Hidden Cause heavy_snow(loc)  drive_hazard(loc)  block_road(loc) heavy_snow(loc)  drive_hazard(loc)  rb_C1(loc) rb_C1(loc) Hidden Cause rb_C1(loc)  block_road(loc)

Introducing Hidden Cause heavy_snow(loc)  drive_hazard(loc)  block_road(loc) heavy_snow(loc)  drive_hazard(loc)  rb_C1(loc) rb_C1(loc)  block_road(loc) accident(loc)  clear_wreck(loc, crew)  block_road(loc) accident(loc)  clear_wreck(loc)  rb_C2(loc, crew) rb_C2(loc, crew) rb_C2(loc,crew)  block_road(loc)

Introducing Reverse Implication block_road(loc)  rb_C1(loc) v (  crew rb_C2(loc, crew)) Explanation 2: accident(loc)  clear_wreck(loc)  rb_C2(loc, crew) Explanation 1: heavy_snow(loc)  clear_wreck(loc)  rb_C1(loc) Multiple causes combined via reverse implication Existential quantification

Low-Prior on Hidden Causes block_road(loc)  rb_C1(loc) v (  crew rb_C2(loc, crew)) Explanation 2: accident(loc)  clear_wreck(loc)  rb_C2(loc, crew) Explanation 1: heavy_snow(loc)  clear_wreck(loc)  rb_C1(loc) Multiple causes combined via reverse implication -w1 rb_C1(loc) -w2 rb_C2(loc, crew)

Hidden Causes: Avoiding Blow-up drive_hazard (Plaza) heavy_snow (Plaza) accident (Plaza) clear_wreck (Tcrew, Plaza) rb_C1 (Plaza) rb_C2 (Tcrew, Plaza) block_road (Tcrew, Plaza) Hidden Cause Model Max clique size = 3 [Singla & Domingos 2011]

Hidden Causes: Avoiding Blow-up drive_hazard (Plaza) heavy_snow (Plaza) accident (Plaza) clear_wreck (Tcrew, Plaza) drive_hazard (Plaza) heavy_snow (Plaza) accident (Plaza) clear_wreck (Tcrew, Plaza) rb_C1 (Plaza) rb_C2 (Tcrew, Plaza) block_road (Tcrew, Plaza) block_road (Tcrew, Plaza) Pair-wise Constraints [Kate & Mooney 2009] Max clique size = 5 Hidden Cause Model Max clique size = 3 [Singla & Domingos 2011]

Second Issue: Ground Network Too Big! Grounding out the full network may be costly Many irrelevant nodes/clauses are created Complicates learning/inference Can focus the grounding (KBMC)

Abductive Model Construction block_road (Plaza) heavy_snow (Plaza) drive_hazard (Plaza) Observation: block_road(Plaza)

Abductive Model Construction Constants: …, Mall, City_Square,... block_road (Mall) heavy_snow (Mall) drive_hazard (Mall) block_road (City_Square) drive_hazard (City_Square) heavy_snow (City_Square) block_road (Plaza) heavy_snow (Plaza) drive_hazard (Plaza) Observation: block_road(Plaza)

Abductive Model Construction Constants: …, Mall, City_Square,... Not a part of abductive proof trees! block_road (City_Square) drive_hazard (City_Square) heavy_snow (City_Square) block_road (Mall) heavy_snow (Mall) drive_hazard (Mall) block_road (Plaza) heavy_snow (Plaza) drive_hazard (Plaza) Observation: block_road(Plaza)

Abductive Model Construction Constants: …, Mall, City_Square,... Not a part of abductive proof trees! block_road (City_Square) drive_hazard (City_Square) heavy_snow (City_Square) block_road (Mall) heavy_snow (Mall) drive_hazard (Mall) block_road (Plaza) heavy_snow (Plaza) drive_hazard (Plaza) Observation: block_road(Plaza) Backward chaining to get proof trees [Stickel 1988]

Abductive Markov Logic [Singla & Domingos 11] Re-encode the MLN rules Introduce reverse implications Construct ground Markov network Use abductive model construction Perform learning and inference

Summary Real world applications Entities and Relations Uncertainty Unifying logical and statistical AI Markov Logic – simple and powerful model Need to do to efficient learning and inference Applications: Abductive Plan Recognition