UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causal Diagrams and the Identification of Causal Effects A presentation of.

Slides:



Advertisements
Similar presentations
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Advertisements

Bayesian Network and Influence Diagram A Guide to Construction And Analysis.
Bayesian Networks, Winter Yoav Haimovitch & Ariel Raviv 1.
Week 21 Basic Set Theory A set is a collection of elements. Use capital letters, A, B, C to denotes sets and small letters a 1, a 2, … to denote the elements.
COUNTING AND PROBABILITY
Parallel Scheduling of Complex DAGs under Uncertainty Grzegorz Malewicz.
Bayesian Networks VISA Hyoungjune Yi. BN – Intro. Introduced by Pearl (1986 ) Resembles human reasoning Causal relationship Decision support system/ Expert.
Introduction of Probabilistic Reasoning and Bayesian Networks
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering An Introduction to Pearl’s Do-Calculus of Intervention Marco Valtorta Department.
COMMENTS ON By Judea Pearl (UCLA). notation 1990’s Artificial Intelligence Hoover.
Causal Networks Denny Borsboom. Overview The causal relation Causality and conditional independence Causal networks Blocking and d-separation Excercise.
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
Chapter 8-3 Markov Random Fields 1. Topics 1. Introduction 1. Undirected Graphical Models 2. Terminology 2. Conditional Independence 3. Factorization.
Bayesian Networks A causal probabilistic network, or Bayesian network,
FUNCTIONS AND MODELS Chapter 1. Preparation for calculus :  The basic ideas concerning functions  Their graphs  Ways of transforming and combining.
1 Introduction to Computability Theory Lecture3: Regular Expressions Prof. Amos Israeli.
PGM 2003/04 Tirgul 3-4 The Bayesian Network Representation.
VECTOR CALCULUS Fundamental Theorem for Line Integrals In this section, we will learn about: The Fundamental Theorem for line integrals and.
Bayesian Network Representation Continued
Inferring Causal Graphs Computing 882 Simon Fraser University Spring 2002.
7 INVERSE FUNCTIONS. The common theme that links the functions of this chapter is:  They occur as pairs of inverse functions. INVERSE FUNCTIONS.
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Uncertainty in Artificial Intelligence Research at USC: Research Presentation.
Copyright © Cengage Learning. All rights reserved.
SIMPLE LINEAR REGRESSION
. DAGs, I-Maps, Factorization, d-Separation, Minimal I-Maps, Bayesian Networks Slides by Nir Friedman.
. PGM 2002/3 – Tirgul6 Approximate Inference: Sampling.
5  Systems of Linear Equations: ✦ An Introduction ✦ Unique Solutions ✦ Underdetermined and Overdetermined Systems  Matrices  Multiplication of Matrices.
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
Copyright © Cengage Learning. All rights reserved. CHAPTER 11 ANALYSIS OF ALGORITHM EFFICIENCY ANALYSIS OF ALGORITHM EFFICIENCY.
Rev.S08 MAC 1140 Module 12 Introduction to Sequences, Counting, The Binomial Theorem, and Mathematical Induction.
Induction and recursion
Copyright © Cengage Learning. All rights reserved. 6 Inverse Functions.
Zvi Kohavi and Niraj K. Jha 1 Memory, Definiteness, and Information Losslessness of Finite Automata.
DECIDABILITY OF PRESBURGER ARITHMETIC USING FINITE AUTOMATA Presented by : Shubha Jain Reference : Paper by Alexandre Boudet and Hubert Comon.
1 7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to.
Introduction Of Tree. Introduction A tree is a non-linear data structure in which items are arranged in sequence. It is used to represent hierarchical.
Advanced Topics in Propositional Logic Chapter 17 Language, Proof and Logic.
11.2 Series In this section, we will learn about: Various types of series. INFINITE SEQUENCES AND SERIES.
Introduction to Bayesian Networks
V13: Causality Aims: (1) understand the causal relationships between the variables of a network (2) interpret a Bayesian network as a causal model whose.
Course files
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS.
INTERVENTIONS AND INFERENCE / REASONING. Causal models  Recall from yesterday:  Represent relevance using graphs  Causal relevance ⇒ DAGs  Quantitative.
Sampling distributions rule of thumb…. Some important points about sample distributions… If we obtain a sample that meets the rules of thumb, then…
1 Bayesian Networks (Directed Acyclic Graphical Models) The situation of a bell that rings whenever the outcome of two coins are equal can not be well.
CS 103 Discrete Structures Lecture 13 Induction and Recursion (1)
Block Ciphers and the Advanced Encryption Standard
1 Use graphs and not pure logic Variables represented by nodes and dependencies by edges. Common in our language: “threads of thoughts”, “lines of reasoning”,
Other Access Control Models
Mathematical Induction Section 5.1. Climbing an Infinite Ladder Suppose we have an infinite ladder: 1.We can reach the first rung of the ladder. 2.If.
Causal Model Ying Nian Wu UCLA Department of Statistics July 13, 2007 IPAM Summer School.
Judea Pearl Computer Science Department UCLA ROBUSTNESS OF CAUSAL CLAIMS.
CAUSAL REASONING FOR DECISION AIDING SYSTEMS COGNITIVE SYSTEMS LABORATORY UCLA Judea Pearl, Mark Hopkins, Blai Bonet, Chen Avin, Ilya Shpitser.
Reasoning Under Uncertainty: Independence and Inference CPSC 322 – Uncertainty 5 Textbook §6.3.1 (and for HMMs) March 25, 2011.
Introduction on Graphic Models
1 BN Semantics 2 – Representation Theorem The revenge of d-separation Graphical Models – Carlos Guestrin Carnegie Mellon University September 17.
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2006 Readings: K&F: 3.1, 3.2, 3.3.
Variable selection in Regression modelling Simon Thornley.
Copyright © Cengage Learning. All rights reserved. 8 9 Correlation and Regression.
Theory of Computational Complexity Probability and Computing Ryosuke Sasanuma Iwama and Ito lab M1.
An Algorithm to Learn the Structure of a Bayesian Network Çiğdem Gündüz Olcay Taner Yıldız Ethem Alpaydın Computer Engineering Taner Bilgiç Industrial.
CHAPTER SIX T HE P ROBABILISTIC M ETHOD M1 Zhang Cong 2011/Nov/28.
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
CS 2750: Machine Learning Directed Graphical Models
What is Probability? Quantification of uncertainty.
A Bayesian Approach to Learning Causal networks
Dependency Models – abstraction of Probability distributions
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
Presentation transcript:

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causal Diagrams and the Identification of Causal Effects A presentation of Chapters 3 and 4 of: Judea Pearl Causality: Models, Reasoning and Inference Cambridge University Press, 2000

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 2 An Example This is example is due to Cochran through Pearl [2000, p.66] “Soil fumigants (X) are used to increase oat crop yields (Y) by controlling the eelworm population (Z).” Last year’s eelworm population (Z 0 ) is an unknown quantity that is strongly correlated with this year’s population. Through laboratory analysis of soil samples, we can determine the eelworm populations before and after the treatments (Z 1 and Z 2 ). Furthermore, we assume that the fumigants do not affect the growth of eelworms surviving the treatment. Instead, eelworm’s growth depends on the population of birds (B), which is correlated with last year’s eelworm population and hence with the treatment itself. Z 3 here represents the eelworm population at the end of the season.” We wish to assess the total effect of the fumigants on yields. But, controlled randomized experiment are unfeasible and Z 0 is unknown. If we got a correct model, can we obtain consistent estimate of the target quantity – the total effect of the fumigants on yields – through observations?

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 3 Graphs as Models of Interventions A causal diagram is a directed acyclic graph G that identifies the causal connections among the variables of interest. The causal reading of a DAG is in terms of functional, rather than probabilistic relationships where pa i are the parents of variable X i in G. u i are mutually independent and represent unobserved factors, including random disturbances. We have the same recursive decomposition as in a Bayesian network:

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 4 External Intervention & Causal Effect The simplest type of external intervention is one in which a single variable, say X i, is forced to take on some fixed value x i. We call such an intervention “atomic” and denote it by do(X i =x i ) or do(x i ) for short. Definition 3.2.1(Causal Effect): Given two disjoint sets of variables, X and Y, the causal effect of X on Y, denoted either as or as, is a function from X to the space of probability distributions on Y. For each realization x of X, gives the probability of Y = y induced by deleting from the model of (3.4) all equations corresponding to variables in X and substituting X=x in the remaining equations.

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 5 Correlation vs. Causation The genotype theory (Fisher, 1958) of smoking and lung cancer: smoking and lung cancer are both effects of a genetic predisposition Three node network X( smoking) and Y( lung cancer) are in lockstep X precedes Y in time (smoke before cancer) But, X does not cause Y, because if we set X, Y does not change: Y only changes according to the value of U (the genotype) XY U

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 6 Effect of Interventions From definition 3.2.1, we have: Theorem 3.2.2(adjustment for Direct causes) Let PA i denote the set of direct causes of variable X i, and let Y be any set of variables disjoint of. The effect of the intervention on Y is given by Where and represent preintervention probabilities.

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 7 Causal Effect Identifiably Definition 3.2.4(Causal Effect Identifiability) The causal effect of X on Y is identifiable from a graph G if the quantity can be computed uniquely from any positive probability of the observed variables – that is, if for every pair of models M 1 and M 2 with the same probability distribution for the set of observed variables (v), i.e. and the same graph, i.e., Theorem Given a causal diagram G of any Markovian model in which a subset V of variables are measured, the causal effect is identifiable whenever, that is, whenever X, Y, and all parents of variables in X are measured. The expression for is then obtained by adjusting for PAx, as in 3.14 A special case of Theorem holds when all variables are assumed to be observed.

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 8 Example of Nonidentifiability The identifiablility of ensures that it is possible to infer the effect of action do(X=x) on Y from passive observations and the causal graph G, which specifies which variables participate in the determination of each variable in the domain. To prove nonidentifiability, it is sufficient to present two sets of structural equations that induce identical distributions over observed variables but have different causal effects. X,Y is observable, U is not. All of them are binary variables. If P(X=0|U) = (0.5,0.5) P(Y=0|X,U) = But we don’t know P(U) When P(U=0) = 0.5, P(Y|X=0) =(.45,.55) When P(U=0) = 0.1, P(Y|X=0) =(.73,.27) So, P(Y|do(X)) is non-identifiable Y=0X =0X= 1 U = U= X Y U

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 9 Example of Nonidentifiability The identifiablility of ensures that it is possible to infer the effect of action do(X=x) on Y from passive observations and the causal graph G, which specifies which variables participate in the determination of each variable in the domain. To prove nonidentifiability, it is sufficient to present two sets of conditional probability tables that induce identical distributions over observed variables but have different causal effects. X,Y is observable, U is not. All of them are binary variables. Let P(X,Y,U) be If P(X=0|U) = (0.5,0.5) P(Y=0|X,U) = But we don’t know P(U) When P(U=0) = 0.5, P(Y|X=0) =(.45,.55) When P(U=0) = 0.1, P(Y|X=0) =(.73,.27) So, P(Y|do(X)) is non-identifiable (U=0, U=1)X =0X= 1 Y =00.5,0.50,0 Y= , 0.5 X Y U

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 10 Fisher vs. the Surgeon General XY U Only X and Y are observable U X YXY U Fisher’s Genotype TheorySurgeon General’s Opinion

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 11 Intervention as surgically modified DAGs

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 12 Different Causal Effects for the Same Observations

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 13 Interventions as Variables The effect of an atomic intervention do(X i =x i ’) is encoded by adding to G a link F i  X i, where F i is a new variable taking values in {do(x i ’),idle}, x i ’ ranges over the domain of X i, and “idle” represents no intervention. Then we define:

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 14 Backdoor Criterion If we can get P’(y|z,x,Fx) = P(y|x,z) and P’(z|Fx) =P(z), then we know is identifiable. Definition 3.3.1(Back-Door) A set of variables Z satisfies that back-door criterion relative to an ordered pair of variables(Xi,Xj) in a DAG G if: (i) no node in Z is a descendant of Xi; and (ii) Z blocks every path between Xi and Xj that contains an arrow into Xi. Similarly, if X and Y are two disjoint subsets of nodes in G, then Z is said to satisfy that back-door criterion relative to (X,Y) if it satisfies that criterion relative to any pair (Xi,Xj) such that Xi  X and Xj  Y. Theorem 3.3.2(Back-Door Adjustment) If a set of variables Z satisfies the back-door criterion relative to (X,Y), then the causal effect of X on Y is identifiable and is given by the formula:

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 15 Keshan Disease To determining the causal effect of Selenium on Keshan Disease, we need to find the variable set Z, called concomitants, which satisfies the back-door criteria. Z={ Region of China} is a answer. Z={Genotype} also, but this variable maybe is not observable. Backdoor Path

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 16 Smoking and the genotype theory Consider the relation between smoking(X) and lung cancer(Y). the tobacco industry has managed to forestall antismoking legislation by arguing that observed correlation between smoking and lung cancer could be explained by some sort of carcinogenic genotype(U) that involves inborn carving for nicotine here, Z is the amount of tar deposited in a person's lungs. Can we get ?

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 17 Effect of Smoking on Lung Cancer in the Presence of Tar Deposits We compute But And, by the backdoor criterion,

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 18 Front-Door Criterion Definition 3.3.3(Front-Door) A set of variables Z is said to satisfy the front-door criterion relative to an ordered pair of variables (X,Y) if: (i) Z intercepts all directed paths from X to Y; (ii) there is no back-door path from X to Z; and (iii) all back-door paths from Z to Y are blocked by X. Theorem (Front-Door Adjustment) if Z satisfies the front-door criterion relative to (X,Y) and if P(x,z) >0, then the causal effect of X on Y is identifiable and is given by the formula:

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 19 Proof of Front_Door Assume u is the parent set of x, from (i), we have The intervention do(x) removes the factor P(x|u) and induces the post intervention distribution. So From (ii) and (iii), we also have: So: Another proof will be given after the three rules.

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 20 Pearl’s Calculus of Interventions Let X,Y and Z be arbitrary disjoint sets of nodes in a causal DAG G. We denote by the graph obtained by deleting from G all arrows pointing to nodes in X. Likewise, we denote by Gx the graph obtained by deleting from G all arrows emerging from nodes in X. to represent the deletion of both incoming and outgoing arrows, we use the notation ( see Figure 3.6 for an illustration). Finally, the expression stands for the probability of Y=y given that X is held constant at x and that (under this condition) Z=z is observed. Each of these inference rules follows from the basic interpretation of the “hat” operator as a replacement of the causal mechanism that connects X to its preaction parents by a new mechanism X =x introduced by the intervening force. The result is a sub model characterized by the subgraph Gx.

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 21 The Three Rules Theorem ( Rules of Do calculus) Let G be the directed acyclic graph associated with a causal model as defined in(3.2), and let P(.) stand for the probability distribution induced by that model. For any disjoint subsets of variables X, Y, Z, and W, we have the following rules. Rule 1 (Insertion/deletion of observations): Rule 1 reaffirms d-separation as a valid test for conditional independence in the distribution resulting from the intervention do(X=x), hence the graph. This rule follows from the fact that deleting equations from the system does not introduce any dependencies among the remaining disturbance terms.

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 22 The Three Rules Rule 2 provides a condition for an external intervention do(Z=z) to have the same effect on Y as the passive observation Z=z, The condition amounts to {X  W} blocking all back-door paths form Z to Y( in ), since retains all (and only ) such paths. Rule 2 ( Action /observation exchange):

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 23 The Three Rules Rule 3 (Insertion/deletion of actions):,where Z(W) is the set of Z-nodes that are not ancestors of any W-node in Consider G’ with intervention arcs F Z  Z added. implies that So, any path from Z to Y that is not blocked by {X,W} in must end in an arrow pointing to Z, otherwise would not hold. In addition, if there is a path from some Z’ of Z to Y that does and in an arrow pointing to Z’, then W must not be a descendant of Z’, otherwise would not hold. Thus the only paths from Y to Z must end in an arrow pointing at Z, and must end in some member of Z(W). Thus,

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 24 Usage of the Three Rules Corollary A causal effect Is identifiable in a model characterized by a graph G if there exists an finite sequence of transformations, each conforming to one of the inference rules in Theorem 3.4.1, that reduces q into a standard (i.e., “hat”-free) probability expression involving observed quantities. Whether Rules 1-3 are sufficient for deriving all identifiable causal effects remains an open question

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 25 The Smoking Example 1, Based on rule 2, we have

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 26 The Smoking Example 2 Note: we can use the same process to prove back-door formula

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 27 The Smoking Example 3 Based on 1,2, and above, we get: This is also a proof of front door formula

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 28 The Smoking Example 4 See 1 and the third formula in 3, we have: 5 See 2

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 29 Some Nonidentifiable Models Why c is not identifiable?

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 30 Some Identifiable Models

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 31 Why They are Identifiable (a), (b) rule 2, (c), (d) back-door (e) front-door (f) (g)

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 32 Why g? Z1 block all directed paths from X to Y Z2 blocks all back- door paths between Y and Z1 in Putting the pieces together, we obtain the claimed result

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 33 Several More Nonidentifiable Models

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 34 Completeness of the Three Rules Completeness is conjectured by Pearl Scheme for finding a counterexample –generate graphs (models) –filter identifiable models by using 3 rules –filter unidentifiable models by using edge subgraph algorithm with figure 3.9 patterns –rewrite P(y|do(x)) using the rules of probability –If some formula for P(y|do(x)) without U exists, a counterexample has been found –If no formula for P(y|do(x)) without U exists, we add the model to the patterns of figure 3.9

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 35 Identify Theorem 4.3.1(Galles and Pearl 1995) Let X and Y denote two singleton variables in a semi_Markovian Model characterized by graph G. A sufficient condition for the identifiability of is that G satisfy one of the following four conditions. 1, there is no back-door path from X to Y in G. 2,There is no directed path from X to Y in G. 3,There exists a set of nodes B that bocks all back-door paths from X to Y so that is identifiable. ( A special case of this condition occurs when B consists entirely of nondescendents of X, in which case reduces immediately to P(b).) 4,There exist sets of nodes Z 1 and Z 2 such that: (i)Z 1 blocks every directed path form X to Y; (ii)Z 2 blocks all back-door paths between Z 1 and Y; (iii)Z 2 blocks all back-door paths between X and Z 1 ; (iv)Z 2 does not activate any back-door paths from X to Y. this condition holds if (i)-(iii) are met and no member of Z 2 is a descendant of X. (A special case of condition 4 occurs when Z 2 =  and there is no back-door path from X to Z 1 or from Z 1 to Y.) Theorem the four conditions of Theorem are necessary for identifiability in do calculus.

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 36 Remarks on Efficiency Theorem if is identifiable for one minimal blocking set B i, then is identifiable for any other minimal set B j. ( for condition 3) Lemma If the query is identifiable and if a set of nodes Z lies on a directed path form X to Y, then the query is identifiable. Theorem let Y 1 and Y 2 be two subsets of nodes such that either (i) no nodes Y 1 are descendants of X or (ii) all nodes Y 1 and Y 2 are descendants of X and all nodes Y 1 are nondescendants of Y 2. A reducing sequence for exists if and only if there are reducing sequences for both and. Theorem If there exists a set Z 1 that meets all of the requirements for Z 1 in condition 4, then the set consisting of the children of X intersected with the ancestors of Y will also meet all of the requirements for Z 1 in condition 4.

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causality: Models, Reasoning and Inference Chapter 3 37 Closed-Form Expression for Control Queries Function: ClosedForm( ). Input:Control query of the form. Output: either a closed-form expression for,in terms of observed variables only, or FAIL when query is not identifiable. 1, if then return P(y). 2, Otherwise, if then return P(y|x). 3, Otherwise, let B=BlockingSet(X,Y), and Pb=ClosedForm( ) ;if Pb!=FAIL, then return. 4. Otherwise, let Z 1 =Children(X)  (Y  Ancestors(Y)), Z 3 =BlockingSet(X,Z 1 ), Z 4 = BlockingSet(Z 1,Y), and Z 2 = Z 3  Z 4 ; if Y  Z 1 and X  Z 2 then return 5, Otherwise return FAIL Steps 3 and 4 invoke the function BlockingSet(X,Y), which selects a set of nodes Z that d-separate X form Y. Step 3 contains a recursive call to the algorithm ClosedForm( ) itself, in order to obtain an expression for causal effect.