Probabilistic Influence & d-separation

Slides:



Advertisements
Similar presentations
Probabilistic Reasoning Bayesian Belief Networks Constructing Bayesian Networks Representing Conditional Distributions Summary.
Advertisements

G53MLE | Machine Learning | Dr Guoping Qiu
Bayesian Networks. Contents Semantics and factorization Reasoning Patterns Flow of Probabilistic Influence.
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Network Optimization Models: Maximum Flow Problems
Classification Neural Networks 1
Perceptron.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Network Optimization Models: Maximum Flow Problems In this handout: The problem statement Solving by linear programming Augmenting path algorithm.
Bayesian belief networks 2. PCA and ICA
Artificial Neural Networks
Data Mining with Neural Networks (HK: Chapter 7.5)
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Approximate Inference 2: Monte Carlo Markov Chain
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
Artificial Neural Networks
1 Approximate Inference 2: Importance Sampling. (Unnormalized) Importance Sampling.
1 Artificial Neural Networks Sanun Srisuk EECP0720 Expert Systems – Artificial Neural Networks.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
Dan Boneh Symmetric Encryption History Crypto. Dan Boneh History David Kahn, “The code breakers” (1996)
V13: Causality Aims: (1) understand the causal relationships between the variables of a network (2) interpret a Bayesian network as a causal model whose.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
CS 4407, Algorithms University College Cork, Gregory M. Provan Network Optimization Models: Maximum Flow Problems In this handout: The problem statement.
Bayesian Network By Zhang Liliang. Key Point Today Intro to Bayesian Network Usage of Bayesian Network Reasoning BN: D-separation.
Chapter 2 Single Layer Feedforward Networks
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Daphne Koller Markov Networks General Gibbs Distribution Probabilistic Graphical Models Representation.
Daphne Koller Bayesian Networks Semantics & Factorization Probabilistic Graphical Models Representation.
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Neural NetworksNN 21 Architecture We consider the architecture: feed- forward NN with one layer It is sufficient to study single layer perceptrons with.
1 Structure Learning (The Good), The Bad, The Ugly Inference Graphical Models – Carlos Guestrin Carnegie Mellon University October 13 th, 2008 Readings:
Daphne Koller Bayesian Networks Semantics & Factorization Probabilistic Graphical Models Representation.
Reasoning Patterns Bayesian Networks Representation Probabilistic
Daphne Koller Introduction Motivation and Overview Probabilistic Graphical Models.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Daphne Koller Independencies Bayesian Networks Probabilistic Graphical Models Representation.
Fall 2004 Backpropagation CS478 - Machine Learning.
Maximum Expected Utility
Chapter 2 Single Layer Feedforward Networks
Context-Specific CPDs
Probabilistic Models for Linear Regression
Classification Neural Networks 1
Neural Networks for Vertex Covering
General Gibbs Distribution
Independence in Markov Networks
Markov Networks Independencies Representation Probabilistic Graphical
General Gibbs Distribution
Preliminaries: Distributions
Luger: Artificial Intelligence, 5th edition
Perceptron as one Type of Linear Discriminants
Bayesian Networks Independencies Representation Probabilistic
Independence in Markov Networks
General Gibbs Distribution
Graduate School of Information Sciences, Tohoku University
I-equivalence Bayesian Networks Representation Probabilistic Graphical
MCMC for PGMs: The Gibbs Chain
Conditional Random Fields
Reasoning Patterns Bayesian Networks Representation Probabilistic
Factorization & Independence
Factorization & Independence
Chapter - 3 Single Layer Percetron
Markov Networks Independencies Representation Probabilistic Graphical
Crypto Encryption Intro to public key.
Independence in Markov Networks
Markov Networks Independencies Representation Probabilistic Graphical
Probabilistic Reasoning
Flow of Probabilistic Influence
Preliminaries: Independence
Variable Elimination Graphical Models – Carlos Guestrin
Seminar on Machine Learning Rada Mihalcea
Presentation transcript:

Probabilistic Influence & d-separation Representation Probabilistic Graphical Models Bayesian Networks Probabilistic Influence & d-separation

When can X influence Y given evidence about Z Intelligence Difficulty Grade Letter SAT pairs

When can X influence Y given evidence about Z Intelligence Difficulty Grade Letter SAT triples

When can X influence Y given evidence about Z Intelligence Difficulty Grade Letter SAT longer trails

Active Trails A trail X1 ─ … ─ Xn is active given Z if:

d-separation Definition: X and Y are d-separated given evidence Z if

Can Flow ≠ Must Flow Degenerate dependency

Can Flow ≠ Must Flow XOR example

Summary Active trail in a graph G  influence might flow in any distribution P that factorizes over G If a trail is active, influence might still not flow in a specific P that factorizes over G Active trail is necessary but not sufficient for probabilistic influence to flow If two nodes are d-separated, they have no active trails, and influence cannot flow in any P

END END END

Suppose q is at a local minimum of a function Suppose q is at a local minimum of a function. What will one iteration of gradient descent do? Leave q unchanged. Change q in a random direction. Move q towards the global minimum of J(q). Decrease q.

Consider the weight update: Which of these is a correct vectorized implementation?

Fig. A corresponds to a=0.01, Fig. B to a=0.1, Fig. C to a=1.