Probabilistic Resolution. Logical reasoning Absolute implications office meeting office talk office pick_book But what if my rules are not absolute?

Slides:



Advertisements
Similar presentations
Markov models and HMMs. Probabilistic Inference P(X,H,E) P(X|E=e) = P(X,E=e) / P(E=e) with P(X,E=e) = sum h P(X, H=h, E=e) P(X,E=e) = sum h,x P(X=x, H=h,
Advertisements

Inconsistent Heuristics
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
Prof. Sin-Min Lee Department of Computer Science
1 COMM 301: Empirical Research in Communication Lecture 15 – Hypothesis Testing Kwan M Lee.
Techniques for Dealing with Hard Problems Backtrack: –Systematically enumerates all potential solutions by continually trying to extend a partial solution.
Copyright © 2002 Cycorp Logical Aspects of Inference Incompleteness in Searching Incompleteness from Resource Bounds and Continuable Searches Inference.
For Monday Read chapter 18, sections 1-2 Homework: –Chapter 14, exercise 8 a-d.
For Monday Finish chapter 14 Homework: –Chapter 13, exercises 8, 15.
Best-First Search: Agendas
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
CPSC 322, Lecture 35Slide 1 Finish VE for Sequential Decisions & Value of Information and Control Computer Science cpsc322, Lecture 35 (Textbook Chpt 9.4)
CPSC 322, Lecture 9Slide 1 Search: Advanced Topics Computer Science cpsc322, Lecture 9 (Textbook Chpt 3.6) January, 22, 2010.
10/24  Exam on 10/26 (Lei Tang and Will Cushing to proctor)
Jie Gao Joint work with Amitabh Basu*, Joseph Mitchell, Girishkumar Stony Brook Distributed Localization using Noisy Distance and Angle Information.
Administrivia/Announcements Project 0 will be taken until Friday 4:30pm –If you don’t submit in the class, you submit to the dept office and ask them.
CPSC 322, Lecture 12Slide 1 CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12 (Textbook Chpt ) January, 29, 2010.
Existential Graphs and Davis-Putnam April 3, 2002 Bram van Heuveln Department of Cognitive Science.
I.1 ii.2 iii.3 iv.4 1+1=. i.1 ii.2 iii.3 iv.4 1+1=
Computational Complexity, Physical Mapping III + Perl CIS 667 March 4, 2004.
Constructing Belief Networks: Summary [[Decide on what sorts of queries you are interested in answering –This in turn dictates what factors to model in.
10/22  Homework 3 returned; solutions posted  Homework 4 socket opened  Project 3 assigned  Mid-term on Wednesday  (Optional) Review session Tuesday.
I.1 ii.2 iii.3 iv.4 1+1=. i.1 ii.2 iii.3 iv.4 1+1=
“Recognizing Integers” September, th Grade Math.
Aspects of Bayesian Inference and Statistical Disclosure Control in Python Duncan Smith Confidentiality and Privacy Group CCSR University of Manchester.
Similar Triangles and other Polygons
An Illustration of Prim's Algorithm
The Normal Distribution The “Bell Curve” The “Normal Curve”
MATH 224 – Discrete Mathematics
University Prep Math Patricia van Donkelaar Course Outcomes By the end of the course,
MATH- CALCULUS I DONE BY- Supatchaya, Piyate,and Sulakshmi.
For Wednesday Read Chapter 11, sections 1-2 Program 2 due.
1 Variable Elimination Graphical Models – Carlos Guestrin Carnegie Mellon University October 11 th, 2006 Readings: K&F: 8.1, 8.2, 8.3,
Daphne Koller Variable Elimination Graph-Based Perspective Probabilistic Graphical Models Inference.
CSE332: Data Abstractions Lecture 24.5: Interlude on Intractability Dan Grossman Spring 2012.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
Percolation Processes Rajmohan Rajaraman Northeastern University, Boston May 2012 Chennai Network Optimization WorkshopPercolation Processes1.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
The Dominating Set and its Parametric Dual  the Dominated Set  Lan Lin prepared for theory group meeting on June 11, 2003.
CPSC 422, Lecture 11Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 2, 2015.
1 Mean Field and Variational Methods finishing off Graphical Models – Carlos Guestrin Carnegie Mellon University November 5 th, 2008 Readings: K&F:
1 Algorithms  Algorithms are simply a list of steps required to solve some particular problem  They are designed as abstractions of processes carried.
Basic Problem Solving Search strategy  Problem can be solved by searching for a solution. An attempt is to transform initial state of a problem into some.
Probabilistic Equational Reasoning Arthur Kantor
© J. Christopher Beck Lecture 16: Local Search.
The minimum cost flow problem. Solving the minimum cost flow problem.
1 Variable Elimination Graphical Models – Carlos Guestrin Carnegie Mellon University October 15 th, 2008 Readings: K&F: 8.1, 8.2, 8.3,
Statistical NLP Winter 2009
CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12
Inference in Bayesian Networks
CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12
The minimum cost flow problem
Recursive Thinking Chapter 9 introduces the technique of recursive programming. As you have seen, recursive programming involves spotting smaller occurrences.
Exact Inference Continued
Recursive Thinking Chapter 9 introduces the technique of recursive programming. As you have seen, recursive programming involves spotting smaller occurrences.
CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
Bayesian Models in Machine Learning
Exponential Functions
Complexity Analysis Variable Elimination Inference Probabilistic
Exact Inference ..
The Bias Variance Tradeoff and Regularization
Notes Solving a System by Elimination
What LIMIT Means Given a function: f(x) = 3x – 5 Describe its parts.
Exact Inference Continued
Backtracking and Branch-and-Bound
Question 34.
Variable Elimination Graphical Models – Carlos Guestrin
Data Mining CSCI 307, Spring 2019 Lecture 11
Presentation transcript:

Probabilistic Resolution

Logical reasoning Absolute implications office meeting office talk office pick_book But what if my rules are not absolute?

Migrating to Probabilities: Graphical Models noisy_office meeting talk pick_book Actually, the original model does not justify the last row

Migrating to Probabilities: Graphical Models noisy_officemeeting talk pick_book

Variable Elimination (VE) noisy_officemeeting talk pick_book

Variable Elimination (VE) noisy_office talk pick_book meeting (noisy_office, pick_book, talk, meeting) (meeting) meeting

Variable Elimination (VE) noisy_office talk pick_book

Variable Elimination (VE) noisy_office pick_book

Variable Elimination (VE) noisy_office

Graphical Models generalize Logic officemeeting talk pick_book

VE generalizes Resolution Resolution A or B B or C A or C A B C AC Variable Elimination There is still an important difference, though.

Story so far Logic uses absolute rules; Probabilistic models can deal with noise, and generalize logic; But...

Logical reasoning ends early office meeting office talk office pick_book... Given evidence meeting, we are done after considering first rule alone.

Ending early in deterministic graphical model Variable Elimination uses all nodes to calculate P(office | meeting) officemeeting talk pick_book

Ending early in deterministic graphical model But if meeting is observed, we dont need to look beyond it office talk pick_book

Ending early in deterministic graphical model We can use smarter algorithms to end early here as well office talk pick_book

Ending early in non-deterministic graphical models Calculating P(noisy_office | meeting) noisy_officemeeting talk pick_book

Ending early in non-deterministic graphical models P(noisy_office | meeting) depends on all nodes noisy_office talk pick_book

Ending early in non-deterministic graphical models noisy_office talk pick_book But we already know P(noisy_office | meeting) [0.99, ] Can we take advantage of this?

Goal A graphical model inference algorithm that derives a bound on solution so far; Ends as soon as bound is good enough; An anytime algorithm.

Probabilistic Resolution Resolution A or B B or C A or C A B C AC Variable Elimination Variable Elimination generalizes Resolution, but neither provides intermediate results nor ends early. Probabilistic Resolution = VE + ending early

Story so far Logic uses absolute rules; Probabilistic models can deal with noise, and generalize logic; Logic ends as soon as possible, graphical models do not; They can if we are willing to use bounds; But how to calculate bounds?

But how to get bounds? QN2N2 N1N1 N4N4 N3N3...

But how to get bounds? QN2N2 N1N1 N4N4 N3N3...

But how to get bounds? QN2N2 N1N1 N4N4 N3N3

QN

QN 1 2 P(Q) N 1 (Q,N) 2 (N) P(Q) N 1 (Q,N) P 2 (N) P(Q) f ( P 2 (N) )

But how to get bounds? QN P(Q) f ( P 2 (N) ) 0101 f P(Q)P 2 (N)

But how to get bounds? QN P(Q)P 2 (N) (0,0,1) (1,0,0) (0,1,0) (0,0,1) (1,0,0) (0,1,0) f P(Q) f ( P 2 (N) )

But how to get bounds? QN P(Q)P 2 (N) (0,0,1) (1,0,0) (0,1,0) (0,0,1) (1,0,0) (0,1,0) f P(Q) f ( P 2 (N) ) bound Infinite number of points! Justify inner shape to be equal to outter one

But how to get bounds? QN P(Q)P 2 (N) (0,0,1) (1,0,0) (0,1,0) (0,0,1) (1,0,0) (0,1,0) f P(Q) f ( P 2 (N) ) Vertices are enough

But how to get bounds? QN P(Q) (0,0,1) (1,0,0) (0,1,0) (0,0,1) (1,0,0) (0,1,0) f P(Q) f ( P 2 (N) ) P 2 (N) No necessary correspondence

But how to get bounds? QN (0,0,1) (1,0,0) (0,1,0) f P(Q) f ( P 2 (N) ) P 2 (N) 01 P(Q) Correspondence would be impossible in this case Make slide with opposite: segment to triangle

But how to get bounds? QN 0101 f P(Q) P(Q) f ( P 2 (N) ) P 2 (N)

Example I QN [0,1][0.36, 0.67] P(Q) f ( P 2 (N) ) P(Q) N (Q,N) P 2 (N) P(Q) (Q,0)P 2 (N=0) + (Q,1)P 2 (N=1) For P 2 (N=0) = 1: P(Q) (Q,0) 1 + (Q,1) 0 P(Q) (Q,0) P(Q=1) = (1,0) / ( (0,0) + (1,0)) P(Q=1) = 0.4 / ( ) = 0.36 For P 2 (N=1) = 1: P(Q) (Q,0) 0 + (Q,1) 1 P(Q) (Q,1) P(Q=1) = (1,1) / ( (0,1) + (1,1)) P(Q=1) = 0.6 / ( ) = 0.67

P 2 (N) Example II QN [0,1][0.5] P(Q) (0,0,1) (1,0,0) (0,1,0) (0,0,1) (0,1,0) f (1,0,0) 0101 f P(Q) P 2 (N)

Example III QN [0,1] P 2 (N)P(Q) (0,0,1) (1,0,0) (0,1,0) (0,0,1) (0,1,0) f (1,0,0) 0101 f P(Q) P 2 (N)

Example IV noisy_officemeeting talk pick_book

Example IV noisy_office talk pick_book

Example IV noisy_office talk pick_book 0.4

Example IV noisy_office pick_book

Example IV noisy_office pick_book 1

Example IV noisy_office

Algorithm Same as Variable Elimination, but update bounds every time a neighbor is eliminated; Bounds always improve at each neighbor elimination; Trade-off between granularity of bound updates (explain granularity) and ordering efficiency.

Complexity Issues Calculating bound is exponential on the size of neighborhood component, so complexity is exponential on largest neighborhood component during execution; This can be larger than tree-width; But finding tree-width is hard anyway.

Preliminary Tests

Conclusions Making Probabilistic Inference more like Logic Inference; Getting an anytime algorithm in the process; Preparing ground for First-order Probabilistic Resolution.