School of Computing Science Simon Fraser University Vancouver, Canada.

Slides:



Advertisements
Similar presentations
Online Max-Margin Weight Learning with Markov Logic Networks Tuyen N. Huynh and Raymond J. Mooney Machine Learning Group Department of Computer Science.
Advertisements

CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
Markov Networks Alan Ritter.
Discriminative Training of Markov Logic Networks
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Discriminative Structure and Parameter.
LEARNING INFLUENCE PROBABILITIES IN SOCIAL NETWORKS Amit Goyal Francesco Bonchi Laks V. S. Lakshmanan University of British Columbia Yahoo! Research University.
Online Max-Margin Weight Learning for Markov Logic Networks Tuyen N. Huynh and Raymond J. Mooney Machine Learning Group Department of Computer Science.
CPSC 322, Lecture 30Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 30 March, 25, 2015 Slide source: from Pedro Domingos UW.
Intelligent Environments1 Computer Science and Engineering University of Texas at Arlington.
Experiments We measured the times(s) and number of expanded nodes to previous heuristic using BFBnB. Dynamic Programming Intuition. All DAGs must have.
Relational Representations Daniel Lowd University of Oregon April 20, 2015.
The IMAP Hybrid Method for Learning Gaussian Bayes Nets Oliver Schulte School of Computing Science Simon Fraser University Vancouver, Canada
Markov Logic Networks Instructor: Pedro Domingos.
A Hierarchy of Independence Assumptions for Multi-Relational Bayes Net Classifiers School of Computing Science Simon Fraser University Vancouver, Canada.
1 Learning Markov Logic Network Structure Via Hypergraph Lifting Stanley Kok Dept. of Computer Science and Eng. University of Washington Seattle, USA Joint.
1 Unsupervised Semantic Parsing Hoifung Poon and Pedro Domingos EMNLP 2009 Best Paper Award Speaker: Hao Xiong.
Random Regression: Example Target Query: P(gender(sam) = F)? Sam is friends with Bob and Anna. Unnormalized Probability: Oliver Schulte, Hassan Khosravi,
Review Markov Logic Networks Mathew Richardson Pedro Domingos Xinran(Sean) Luo, u
Speeding Up Inference in Markov Logic Networks by Preprocessing to Reduce the Size of the Resulting Grounded Network Jude Shavlik Sriraam Natarajan Computer.
Efficient Weight Learning for Markov Logic Networks Daniel Lowd University of Washington (Joint work with Pedro Domingos)
Markov Logic: A Unifying Framework for Statistical Relational Learning Pedro Domingos Matthew Richardson
Speaker:Benedict Fehringer Seminar:Probabilistic Models for Information Extraction by Dr. Martin Theobald and Maximilian Dylla Based on Richards, M., and.
Learning Markov Network Structure with Decision Trees Daniel Lowd University of Oregon Jesse Davis Katholieke Universiteit Leuven Joint work with:
Modelling Relational Statistics With Bayes Nets School of Computing Science Simon Fraser University Vancouver, Canada Tianxiang Gao Yuke Zhu.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
A Tractable Pseudo-Likelihood for Bayes Nets Applied To Relational Data Oliver Schulte School of Computing Science Simon Fraser University Vancouver, Canada.
Causal Modelling for Relational Data Oliver Schulte School of Computing Science Simon Fraser University Vancouver, Canada.
1 Statistical Predicate Invention Stanley Kok Dept. of Computer Science and Eng. University of Washington Joint work with Pedro Domingos.
CSE 574 – Artificial Intelligence II Statistical Relational Learning Instructor: Pedro Domingos.
Relational Models. CSE 515 in One Slide We will learn to: Put probability distributions on everything Learn them from data Do inference with them.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1 Learning the Structure of Markov Logic Networks Stanley Kok & Pedro Domingos Dept. of Computer Science and Eng. University of Washington.
Statistical Relational Learning Pedro Domingos Dept. Computer Science & Eng. University of Washington.
Boosting Markov Logic Networks
A simple method for multi-relational outlier detection Sarah Riahi and Oliver Schulte School of Computing Science Simon Fraser University Vancouver, Canada.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Charu Aggarwal + * Department of Computer Science, University of Texas at Dallas + IBM T. J. Watson.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Michael Baron + * Department of Computer Science, University of Texas at Dallas + Department of Mathematical.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Querying Structured Text in an XML Database By Xuemei Luo.
第十讲 概率图模型导论 Chapter 10 Introduction to Probabilistic Graphical Models
Report on Intrusion Detection and Data Fusion By Ganesh Godavari.
Biswanath Panda, Mirek Riedewald, Daniel Fink ICDE Conference 2010 The Model-Summary Problem and a Solution for Trees 1.
Markov Logic And other SRL Approaches
Transfer in Reinforcement Learning via Markov Logic Networks Lisa Torrey, Jude Shavlik, Sriraam Natarajan, Pavan Kuppili, Trevor Walker University of Wisconsin-Madison,
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 5.
Markov Logic and Deep Networks Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
Inference Complexity As Learning Bias Daniel Lowd Dept. of Computer and Information Science University of Oregon Joint work with Pedro Domingos.
Learning Linear Causal Models Oksana Kohutyuk ComS 673 Spring 2005 Department of Computer Science Iowa State University.
Tuffy Scaling up Statistical Inference in Markov Logic using an RDBMS
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
Slides for “Data Mining” by I. H. Witten and E. Frank.
Conditional Random Fields for ASR Jeremy Morris July 25, 2006.
Challenge Paper: Marginal Probabilities for Instances and Classes Oliver Schulte School of Computing Science Simon Fraser University Vancouver, Canada.
Bayesian Optimization Algorithm, Decision Graphs, and Occam’s Razor Martin Pelikan, David E. Goldberg, and Kumara Sastry IlliGAL Report No May.
Happy Mittal (Joint work with Prasoon Goyal, Parag Singla and Vibhav Gogate) IIT Delhi New Rules for Domain Independent Lifted.
Introduction on Graphic Models
Progress Report ekker. Problem Definition In cases such as object recognition, we can not include all possible objects for training. So transfer learning.
Learning Bayesian Networks for Complex Relational Data
An Introduction to Markov Logic Networks in Knowledge Bases
Learning Deep Generative Models by Ruslan Salakhutdinov
Boosted Augmented Naive Bayes. Efficient discriminative learning of
Markov Logic Networks for NLP CSCI-GA.2591
CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12
Fast Learning of Relational Dependency Networks
Markov Networks.
Discriminative Frequent Pattern Analysis for Effective Classification
Bucket Renormalization for Approximate Inference
Sofia Pediaditaki and Mahesh Marina University of Edinburgh
Presentation transcript:

School of Computing Science Simon Fraser University Vancouver, Canada

2 Learning Markov Logic Networks Outline Markov Logic Networks (MLNs) and motivation for this research Parameterized Bayes nets (PBNs) From PBN to MLN Learn-and-Join Algorithm for PBNs. Empirical Evaluation

Markov Logic Networks ( Domingos and Richardson ML 2006 ) Learning Markov Logic Networks 3 A logical KB is a set of hard constraints on the set of possible worlds MLNs make them soft constraints: When a world violates a formula, It becomes less probable, not impossible A Markov Logic Network (MLN) is a set of pairs (F, w) where F is a formula in first-order logic w is a real number 2.3If a student is intelligent then he is well ranked 0.7If a student has an intelligent friend then he is intelligent

Why MLNs are important Learning Markov Logic Networks 4 MLNs are proposed as a unifying framework for Statistical relational learning. Their authors show how other approaches are special cases of MLNs. They are popular Learning the structure of Markov logic networks Discriminative training of Markov logic networks Efficient weight learning for Markov logic networks Bottom-up learning of Markov logic network structure Mapping and revising Markov logic networks for transfer learning Discriminative structure and parameter learning for Markov logic networks Entity resolution with markov logic Event modeling and recognition using markov logic networks Hybrid markov logic networks Learning Markov logic network structure via hypergraph lifting Improving the accuracy and efficiency of map inference for markov logic

Limitation Learning Markov Logic Networks 5 Structure learning in MLNs is mostly ILP(Inductive Logic Programming) based The complexity of search space is exponential in the number of predicates For datasets with many descriptive attributes current MLN learning algorithms are infeasible as they either never terminate or are very slow

Parametrized Bayes Nets (Poole UAI 2003) Learning Markov Logic Networks 6 A functor is a function symbol with typed variables f(X), g(X,Y), R(X,Y). A PBN is a BN whose nodes are functors. We use PBNs with variables only. Not intelligence(Jack)

Overview Learning Markov Logic Networks 7

From PBN to FOL formula Learning Markov Logic Networks 8 Parameterized Bayes Nets(PBNs) can be converted to a set of first order formula easily. 1. Moralize PBN (marry parents, drop arrows). 2. For every CP-table value in PBN, add a formula ranking (S 1, R 1 ), intelligence(S 1, I 1 ) ranking (S 1, R 1 ), intelligence(S 1, I 2 ) ranking (S 1, R 2 ), intelligence(S 1, I 1 ) ranking (S 1, R 2 ), intelligence(S 1, I 2 ) Int =I 1 P1P1 P2P2 Int =I 2 P3P3 P4P4

Learning PBN Structure Required: single-table BN learner L. Takes as input Single data table. A set of edge constraints required edges A set of edge constraints forbidden edges Nodes: Descriptive attributes (e.g. intelligence(S)) Boolean relationship nodes (e.g., Registered(S,C)). Edges Learning correlations between attributes Learning Markov Logic Networks 9

Phase 1: Entity tables Learning Markov Logic Networks 10 BN learner L

Phase 2: relationship tables Learning Markov Logic Networks 11 S.NameC.numbergradesatisfactionintelligencerankingratingdifficultyPopularityTeach-ability Jack101A …. …………………… BN learner L

Phase 3: add Boolean relationship indicator variables 12

Datasets University Movielens Mutagenesis 13 Learning Markov Logic Networks

Systems Learning Markov Logic Networks 14 Moralized Bayes Nets(MBN) is a Parameterized Bayes nets (PBNs) converted into a Markov logic network. LHL(current implementation of structure learning in Alchemy) Const_LHL: Following the data reformatting used by Kok and Domingos Salary(Student, salary_value) => Salary_high(student) Salary_low(student) Salary_medium(student)

Evaluation Plan 15 Learning Markov Logic Networks

Evaluation Metrics(Default) Running time Accuracy How accurate our prediction is Conditional Log Likelihood (CLL) How confident we are with the prediction Area Under Curve (AUC) Avoid false negatives 16 Learning Markov Logic Networks

Running time Time in Minutes. NT = did not terminate. X+Y = PBN structure learning + MLN parameter learning 17 Learning Markov Logic Networks MBN

Accuracy 18 Learning Markov Logic Networks

Conditional Log likelihood 19 Learning Markov Logic Networks

Area Under Curve 20 Learning Markov Logic Networks

Future Work: Parameter Learning 21 Learning Markov Logic Networks

Summary Learning Markov Logic Networks 22 Key idea: learn directed model, convert to undirected to avoid cycles. New efficient structure learning algorithm for Parametrized Bayes Nets.  Fast and scalable (e.g., 5 min vs. 21 hr).  Substantial Improvements in all default Evaluation Metrics

Thank you! Learning Markov Logic Networks 23 Any questions?

Learning PBN Structure 24 Required: single-table BN learner L. Takes as input (T,RE,FE): Single data table. A set of edge constraints (forbidden/required edges). Nodes: Descriptive attributes (e.g. intelligence(S)) Boolean relationship nodes (e.g., Registered(S,C)). Required: single-table BN learner L. Takes as input (T,RE,FE): Single data table. A set of edge constraints (forbidden/required edges). Nodes: Descriptive attributes (e.g. intelligence(S)) Boolean relationship nodes (e.g., Registered(S,C)). 1.RequiredEdges, ForbiddenEdges := emptyset. 2.For each entity table E i : a)Apply L to E i to obtain BN G i. For two attributes X,Y from E i, b)If X → Y in G i, then RequiredEdges += X → Y. c)If X → Y not in G i, then ForbiddenEdges += X → Y. 3.For each relationship table join of size s = 1,..k a)Compute Rtable join, join with entity tables := J i. b)Apply L to (J i, RE, FE) to obtain BN G i. c)Derive additional edge constraints from G i. 4.Add relationship indicators: If edge X → Y was added when analyzing join R 1 join R 2 … join R m, add edges R i → Y. 1.RequiredEdges, ForbiddenEdges := emptyset. 2.For each entity table E i : a)Apply L to E i to obtain BN G i. For two attributes X,Y from E i, b)If X → Y in G i, then RequiredEdges += X → Y. c)If X → Y not in G i, then ForbiddenEdges += X → Y. 3.For each relationship table join of size s = 1,..k a)Compute Rtable join, join with entity tables := J i. b)Apply L to (J i, RE, FE) to obtain BN G i. c)Derive additional edge constraints from G i. 4.Add relationship indicators: If edge X → Y was added when analyzing join R 1 join R 2 … join R m, add edges R i → Y.

Restrictions  Learn-and-Join learns dependencies among attributes, not dependencies among relationships.  The structure is limited to certain patterns  Only works with many descriptive attributes  Parameter learning still a bottleneck. Learning Markov Logic Networks 25

Inheriting Edge Constraints From Subtables Learning Markov Logic Networks 26 Intuition: Evaluate dependencies on the smallest possible join. Statistical Motivation: Statistics can change between tables, e.g. distribution of students’ age may differ in Student table from Registration table. Computational Motivation: as larger join tables are formed, many edges need not be considered → fast learning.