Performing Bayesian Inference by Weighted Model Counting Tian Sang, Paul Beame, and Henry Kautz Department of Computer Science & Engineering University.

Slides:



Advertisements
Similar presentations
Adnan Darwiche Computer Science Department UCLA
Advertisements

Lower Bounds for Exact Model Counting and Applications in Probabilistic Databases Paul Beame Jerry Li Sudeepa Roy Dan Suciu University of Washington.
Time-Space Tradeoffs in Resolution: Superpolynomial Lower Bounds for Superlinear Space Chris Beck Princeton University Joint work with Paul Beame & Russell.
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
“Using Weighted MAX-SAT Engines to Solve MPE” -- by James D. Park Shuo (Olivia) Yang.
Using Problem Structure for Efficient Clause Learning Ashish Sabharwal, Paul Beame, Henry Kautz University of Washington, Seattle April 23, 2003.
Daniel Kroening and Ofer Strichman 1 Decision Procedures An Algorithmic Point of View SAT.
1/30 SAT Solver Changki PSWLAB SAT Solver Daniel Kroening, Ofer Strichman.
Phase Transitions of PP-Complete Satisfiability Problems D. Bailey, V. Dalmau, Ph.G. Kolaitis Computer Science Department UC Santa Cruz.
Willis Lemasters Grant Conklin. Searching a tree recursively one branch at a time, abandoning any branch which does not satisfy the search constraints.
Methods for SAT- a Survey Robert Glaubius CSCE 976 May 6, 2002.
1 Boolean Satisfiability in Electronic Design Automation (EDA ) By Kunal P. Ganeshpure.
Ryan Kinworthy 2/26/20031 Chapter 7- Local Search part 1 Ryan Kinworthy CSCE Advanced Constraint Processing.
Algorithms in Exponential Time. Outline Backtracking Local Search Randomization: Reducing to a Polynomial-Time Case Randomization: Permuting the Evaluation.
1 Towards Efficient Sampling: Exploiting Random Walk Strategy Wei Wei, Jordan Erenrich, and Bart Selman.
State-of-the-art in SAT solvers
1 Sampling, Counting, and Probabilistic Inference Wei joint work with Bart Selman.
GRASP SAT solver Presented by Constantinos Bartzis Slides borrowed from Pankaj Chauhan J. Marques-Silva and K. Sakallah.
Constructing Belief Networks: Summary [[Decide on what sorts of queries you are interested in answering –This in turn dictates what factors to model in.
1 Understanding the Power of Clause Learning Ashish Sabharwal, Paul Beame, Henry Kautz University of Washington, Seattle IJCAI ConferenceAug 14, 2003.
SAT Algorithms in EDA Applications Mukul R. Prasad Dept. of Electrical Engineering & Computer Sciences University of California-Berkeley EE219B Seminar.
CS 188: Artificial Intelligence Spring 2007 Lecture 14: Bayes Nets III 3/1/2007 Srini Narayanan – ICSI and UC Berkeley.
1 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Satisfiability (Reading R&N: Chapter 7)
CS 188: Artificial Intelligence Fall 2006 Lecture 17: Bayes Nets III 10/26/2006 Dan Klein – UC Berkeley.
1 Message Passing and Local Heuristics as Decimation Strategies for Satisfiability Lukas Kroc, Ashish Sabharwal, Bart Selman (presented by Sebastian Brand)
Solution Counting Methods for Combinatorial Problems Ashish Sabharwal [ Cornell University] Based on joint work with: Carla Gomes, Willem-Jan van Hoeve,
Caching in Backtracking Search Fahiem Bacchus University of Toronto.
Satisfiability Introduction to Artificial Intelligence COS302 Michael L. Littman Fall 2001.
Ten Challenges Redux: Recent Progress in Propositional Reasoning & Search A Biased Random Walk Henry Kautz University of Washington.
Boolean Satisfiability and SAT Solvers
1 MCMC Style Sampling / Counting for SAT Can we extend SAT/CSP techniques to solve harder counting/sampling problems? Such an extension would lead us to.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Michael Baron + * Department of Computer Science, University of Texas at Dallas + Department of Mathematical.
INTRODUCTION TO ARTIFICIAL INTELLIGENCE COS302 MICHAEL L. LITTMAN FALL 2001 Satisfiability.
Solvers for the Problem of Boolean Satisfiability (SAT) Will Klieber Aug 31, 2011 TexPoint fonts used in EMF. Read the TexPoint manual before you.
1 Agenda Modeling problems in Propositional Logic SAT basics Decision heuristics Non-chronological Backtracking Learning with Conflict Clauses SAT and.
Heavy-Tailed Phenomena in Satisfiability and Constraint Satisfaction Problems by Carla P. Gomes, Bart Selman, Nuno Crato and henry Kautz Presented by Yunho.
Toward a Universal Inference Engine Henry Kautz University of Washington With Fahiem Bacchus, Paul Beame, Toni Pitassi, Ashish Sabharwal, & Tian Sang.
Inference Complexity As Learning Bias Daniel Lowd Dept. of Computer and Information Science University of Oregon Joint work with Pedro Domingos.
Survey Propagation. Outline Survey Propagation: an algorithm for satisfiability 1 – Warning Propagation – Belief Propagation – Survey Propagation Survey.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
Variable and Value Ordering for MPE Search Sajjad Siddiqi and Jinbo Huang.
Conformant Probabilistic Planning via CSPs ICAPS-2003 Nathanael Hyafil & Fahiem Bacchus University of Toronto.
Combining Component Caching and Clause Learning for Effective Model Counting Tian Sang University of Washington Fahiem Bacchus (U Toronto), Paul Beame.
On the Relation between SAT and BDDs for Equivalence Checking Sherief Reda Rolf Drechsler Alex Orailoglu Computer Science & Engineering Dept. University.
1 The Wumpus Game StenchBreeze Stench Gold Breeze StenchBreeze Start  Breeze.
SAT 2009 Ashish Sabharwal Backdoors in the Context of Learning (short paper) Bistra Dilkina, Carla P. Gomes, Ashish Sabharwal Cornell University SAT-09.
Review of Propositional Logic Syntax
© Copyright 2008 STI INNSBRUCK Intelligent Systems Propositional Logic.
Nikolaj Bjørner Microsoft Research DTU Winter course January 2 nd 2012 Organized by Flemming Nielson & Hanne Riis Nielson.
Heuristics for Efficient SAT Solving As implemented in GRASP, Chaff and GSAT.
SAT Solving As implemented in - DPLL solvers: GRASP, Chaff and
AAAI of 20 Deconstructing Planning as Satisfiability Henry Kautz University of Rochester in collaboration with Bart Selman and Jöerg Hoffmann.
Local Search Methods for SAT Geoffrey Levine March 11, 2004.
1 Boolean Satisfiability (SAT) Class Presentation By Girish Paladugu.
1 Structure Learning (The Good), The Bad, The Ugly Inference Graphical Models – Carlos Guestrin Carnegie Mellon University October 13 th, 2008 Readings:
A COURSE ON PROBABILISTIC DATABASES Dan Suciu University of Washington June, 2014Probabilistic Databases - Dan Suciu 1.
CSC 444 Presentation (11/30/10) Bayesian reasoning using MAX-SAT and model-counting Shuo Yang Karl L. Stratos (a.k.a. Jang Sun Lee)
REU 2007-ParSat: A Parallel SAT Solver Christopher Earl, Mentor: Dr. Hao Zheng Department of Computer Science & Engineering Introduction Results and Conclusions.
Heuristics for Efficient SAT Solving As implemented in GRASP, Chaff and GSAT.
Hybrid BDD and All-SAT Method for Model Checking
Inference and search for the propositional satisfiability problem
Planning as Satisfiability
Inference in Bayesian Networks
Solving MAP Exactly by Searching on Compiled Arithmetic Circuits
A New Algorithm for Computing Upper Bounds for Functional EmajSAT
Encoding CNFs to Enhance Component Analysis
Functional Treewidth:
Hidden Markov Models Part 2: Algorithms
ECE 667 Synthesis and Verification of Digital Circuits
Decision Procedures An Algorithmic Point of View
Presentation transcript:

Performing Bayesian Inference by Weighted Model Counting Tian Sang, Paul Beame, and Henry Kautz Department of Computer Science & Engineering University of Washington University of Washington Seattle, WA

Goal ► Extend success of “compilation to SAT” work for NP-complete problems to “compilation to #SAT” for #P-complete problems  Leverage rapid advances in SAT technology  Example: Computing permanent of a 0/1 matrix  Inference in Bayesian networks (Roth 1996, Dechter 1999) ► Provide practical reasoning tool ► Demonstrate relationship between #SAT and conditioning algorithms  In particular: compilation to DNNF (Darwiche 2002, 2004)

Contributions ► Simple encoding of Bayesian networks into weighted model counting ► Techniques for extending state-of-the-art SAT algorithms for efficient weighted model counting ► Evaluation on computationally challenging domains  Outperforms join-tree methods on problems with high tree-width  Competitive with best conditioning methods on problems with high degree of determinism

Outline ► Model counting ► Encoding Bayesian networks ► Related Bayesian inference algorithms ► Experiments  Grid networks  Plan recognition ► Conclusion

SAT and #SAT ► Given a CNF formula,  SAT: find a satisfying assignment n  #SAT: count satisfying assignments ► Example: (x  y)  (y   z)  5 models: (0,1,0), (0,1,1), (1,1,0), (1,1,1), (1, 0, 0)  Equivalently: satisfying probability = 5/2 3 ► Probability that formula is satisfied by a random truth assignment ► Can modify Davis-Putnam-Logemann-Loveland to calculate this value

DPLL for SAT DPLL(F) if F is empty, return 1 if F contains an empty clause, return 0 else choose a variable x to branch return (DPLL(F| x=1 ) V DPLL(F| x=0 )) return (DPLL(F| x=1 ) V DPLL(F| x=0 )) #DPLL for #SAT #DPLL(F)// computes satisfying probability of F if F is empty, return 1 if F contains an empty clause, return 0 else choose a variable x to branch return 0.5*#DPLL(F| x=1 ) + 0.5*#DPLL(F| x=0 ) return 0.5*#DPLL(F| x=1 ) + 0.5*#DPLL(F| x=0 )

Weighted Model Counting ► Each literal has a weight  Weight of a model = Product of weight of its literals  Weight of a formula = Sum of weight of its models WMC(F) if F is empty, return 1 if F contains an empty clause, return 0 else choose a variable x to branch return weight(x) * WMC(F| x=1 ) + return weight(x) * WMC(F| x=1 ) + weight(  x) * WMC(F| x=0 ) weight(  x) * WMC(F| x=0 )

Cachet ► State of the art model counting program (Sang, Bacchus, Beame, Kautz, & Pitassi 2004) ► Key innovation: sound integration of component caching and clause learning  Component analysis (Bayardo & Pehoushek 2000 ): if formulas C 1 and C share no variables,  Component analysis (Bayardo & Pehoushek 2000 ): if formulas C 1 and C 2 share no variables, BWMC (C  C) = BWMC (C) * BWMC (C) BWMC (C 1  C 2 ) = BWMC (C 1 ) * BWMC (C 2 )  Caching (Majercik & Littman 1998; Darwiche 2002; Bacchus, Dalmao, & Pitassi 2003; Beame, Impagliazzo, Pitassi, & Segerland 2003) : save and reuse values of internal nodes of search tree  Clause learning (Marquis-Silva 1996; Bayardo & Shrag 1997; Zhang, Madigan, Moskewicz, & Malik 2001 ) : analyze reason for backtracking, store as a new clause

Cachet ► State of the art model counting program (Sang, Bacchus, Beame, Kautz, & Pitassi 2004) ► Key innovation: sound integration of component caching and clause learning  Naïve combination of all three techniques is unsound  Can resolve by careful cache management (Sang, Bacchus, Beame, Kautz, & Pitassi 2004)  New branching strategy (VSADS) optimized for counting (Sang, Beame, & Kautz SAT-2005)

Computing All Marginals ► Task: In one counting pass,  Compute number of models in which each literal is true  Equivalently: compute marginal satisfying probabilities ► Approach  Each recursion computes a vector of marginals  At branch point: compute left and right vectors, combine with vector sum  Cache vectors, not just counts ► Reasonable overhead: 10% - 40% slower than counting

Encoding Bayesian Networks to Weighted Model Counting A B A AAAA BBBBB0.1A

A B A AAAA BBBBB0.1A Chance variable P added with weight(P)=0.2

Encoding Bayesian Networks to Weighted Model Counting A B A AAAA BBBBB0.1A  P)=0.8 and weight(  P)=0.8

Encoding Bayesian Networks to Weighted Model Counting A B A AAAA BBBBB0.1A Chance variable Q added with weight(Q)=0.6

Encoding Bayesian Networks to Weighted Model Counting A B A AAAA BBBBB0.1A  Q)=0.4 and weight(  Q)=0.4

Encoding Bayesian Networks to Weighted Model Counting A B A AAAA BBBBB0.1Aw(A)=0.1 w(  A)=0.9 w(P)=0.2 w(  P)=0.8 w(Q)=0.6 w(  Q)=0.4 w(B)=1.0 w(  B)=1.0

Main Theorem ► Let:  F = a weighted CNF encoding of a Bayes net  E = an arbitrary CNF formula, the evidence  Q = an arbitrary CNF formula, the query ► Then:

Exact Bayesian Inference Algorithms ► Junction tree algorithm (Shenoy & Shafer 1990)  Most widely used approach  Data structure grows exponentially large in tree-width of underlying graph ► To handle high tree-width, researchers developed conditioning algorithms, e.g.:  Recursive conditioning (Darwiche 2001)  Value elimination (Bacchus, Dalmao, Pitassi 2003)  Compilation to d-DNNF (Darwiche 2002; Chavira, Darwiche, Jaeger 2004; Darwiche 2004) ► These algorithms become similar to DPLL...

Techniques Method Cache index Cache value Branching heuristic Clause learning? Weighted Model Counting componentprobabilitydynamic Recursive Conditioning partial assignment probabilitystatic Value Elimination dependency set probability semi- dynamic Compiling to d-DNNF residual formula d-DNNF semi- dynamic

Experiments ► Our benchmarks: Grid, Plan Recognition  Junction tree - Netica  Recursive conditioning – SamIam  Value elimination – Valelim  Weighted model counting – Cachet ► ISCAS-85 and SATLIB benchmarks  Compilation to d-DNNF – timings from (Darwiche 2004)  Weighted model counting - Cachet

Experiments: Grid Networks S T ► CPT’s are set randomly. ► A fraction of the nodes are deterministic, specified as a parameter ratio. ► T is the query node

Results of ratio=0.5 Size Junction Tree Recursive Conditioning Value Elimination Weighted Model Counting 10* * * *16X104>20, *18X2,126X13, problems of each size, X=memory out or time out

Results of ratio=0.75 Size Junction Tree Recursive Conditioning Value Elimination Weighted Model Counting 12* * *16> *18X1,751>1, *20X>24,026>94, *22XXX1,300 24*24XXX4,998

Results of ratio=0.9 Size Junction Tree Recursive Conditioning Value Elimination Weighted Model Counting 16* *18X *20X>44, *24XX *26XX>8, *30XXX108

Plan Recognition ► Task:  Given a planning domain described by STRIPS operators, initial and goal states, and time horizon  Infer the marginal probabilities of each action ► Abstraction of strategic plan recognition: We know enemy’s capabilities and goals, what will it do? ► Modified Blackbox planning system (Kautz & Selman 1999) to create instances

problemvariables Junction Tree Recursive Conditioning Value Elimination Weighted Model Counting 4-step step tire-1352XX tire-2550XX tire-3577XX tire-4812XX251.1 log-1939XX log-21337XXX7.9 log-31413XXX9.7 log-42303XXX65

ISCAS/SATLIB Benchmarks Benchmarks reported in (Darwiche 2004) Compiling to d-DNNF Weighted Model Counting uf200 (100 instances) 137 flat200 (100 instances) 508 c c c ,506 c ,057 c ,855

Summary ► Bayesian inference by translation to model counting is competitive with best known algorithms for problems with  High tree-width  High degree of determinism ► Recent conditioning algorithms already make use of important SAT techniques  Most striking: compilation to d-DNNF ► Translation approach makes it possible to quickly exploit future SAT algorithms and implementations