Presentation is loading. Please wait.

Presentation is loading. Please wait.

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Conflicts in Bayesian Networks January 23, 2007 Marco Valtorta

Similar presentations


Presentation on theme: "UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Conflicts in Bayesian Networks January 23, 2007 Marco Valtorta"— Presentation transcript:

1 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Conflicts in Bayesian Networks January 23, 2007 Marco Valtorta mgv@cse.sc.edu

2 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Example: Case Study #4 Bayesian Network Fragment Matching 1) Report Date: 1 April, 2003. FBI: Abdul Ramazi is the owner of the Select Gourmet Foods shop in Springfield Mall. Springfield, VA. (Phone number 703-659.2317). First Union National Bank lists Select Gourmet Foods as holding account number 1070173749003. Six checks totaling $35,000 have been deposited in this account in the past four months and are recorded as having been drawn on accounts at the Pyramid Bank of Cairo, Egypt and the Central Bank of Dubai, United Arab Emirates. Both of these banks have just been listed as possible conduits in money laundering schemes. Partially- Instantiated Bayesian Network Fragment ….. …. BN Fragment Repository

3 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Example: Case Study #4 Bayesian Network Fragment Composition..... + Fragments Situation-Specific Scenario

4 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Value of Information An item of information is useful if acquiring it leads to a better decision, that is, to a more useful action An item of information is useless if the actions that are taken after acquiring it are no more useful than before acquiring it In particular, information is useless if the actions that are taken after acquiring it are the same as before acquiring it In the absence of a detailed model of the utility of actions, the decrease in uncertainty about a variable of interest is taken to be a proxy for the increase in utility: the best item of information to acquire is the one that reduces the most the uncertainty about a variable of interest Since the value of the new item of information is not known, we average over its possible values Uncertainty is measured by entropy. Reduction in uncertainty is measured by reduction in entropy

5 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Example: Case Study #4 Computing Value of Information and Surprise This is the output of the VOI program on a situation-specific scenario for Case Study #4 (Sign of the Crescent). Variable Travel (which represents suspicious travel) is significant for determining the state of variable Suspect (whether Ramazi is a terrorist), even when it is already known that Ramazi has performed suspicious banking transactions. Ramazi performed illegal banking transactions Is Ramazi a terrorist? Would it help to know whether he traveled to sensitive locations? Yes

6 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Value of Information: Formal Definition Let V be a variable whose value affects the actions to be taken by an analyst. For example, V indicates whether a bomb is placed on a particular airliner Let p(v) be the probability that variable V has value v. The entropy of V is: Let T be a variable whose value we may acquire (by expending resources). For example, T indicates whether a passenger is a known terrorist. The entropy of V given that T has value t is: The expected entropy of V given T is: The value of information is:

7 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Surprise Detection Surprise is the situation in which evidence (a set of findings) and a situation-specific scenario are incompatible Since situation-specific scenarios are Bayesian networks, it is very unusual for an outright inconsistency to occur In some cases, however, the evidence is very unlikely in a given scenario; this may be because a rare case has been found, or because the scenario cannot explain the evidence To distinguish these two situations, we compare the probability of the evidence in the situation-specific scenario to the probability of the evidence in a scenario in which all events are probabilistically independent and occur with the same prior probability as in the situation-specific scenario

8 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering The VALUE OF INFORMATION of the test node C for the target node A is 0.0 Parsing the XMLBIF file 'ssn.xml'... done! PROBABILITY FOR JOINT FINDINGS = 5.0E-4 Prior probability for NODE: Suspicious Person=yes is 0.01 Prior probability for NODE: Unusual Activities=yes is 0.0656 Prior probability for NODE: Stolen Weapons=yes is 0.05 PROBABILITY FOR INDIVIDUAL FINDINGS = 3.28E-5 No conflict was detected. This shows the output of the surprise detection program. In this case, the user is informed that no conflict is detected, i.e., the scenario is likely to be a good interpretive model for the evidence received Example: Case Study #4 Computing Surprise

9 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Surprise Detection: Formal Definition Let the evidence be a set of findings: The probability of the evidence in the situation-specific scenario is where is the distribution represented in the situation-specific scenario The probability of the evidence in the model in which all variables are independent is The evidence is surprising if The conflict index is defined as The probability under that is greater than is Proof [Laskey, 1991]: If the conflict index is high, it is unlikely that the findings could have been generated by sampling the situation-specific scenario It is reasonable to inform the analyst that no good explanatory model of the findings exists, and we are in the presence of a novel or surprising situation

10 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering The Independent Straw Model In the absence of conflict, the joint probability of all evidence variables is greater than the product of the probabilities of each evidence variable. This is normally the case, because P(x|y) > P(x), and P(x,y) = P(x|y)P(y).

11 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Straw Models in Diagnosis A bipartite straw model is obtained by the elimination of some variables from a given model. In diagnosis by heuristic classification, one can divide variables into three sets: Target, Evidence, and Other

12 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering How to Compute the Conflict Index (I) The marginal probability of each finding is the normal result of any probability computation algorithm

13 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering How to Compute the Conflict Index (II) The probability of the evidence is a bi-product of probability update computed using the variable elimination or junction tree algorithms

14 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering P(e) from the Variable Elimination Algorithm Bucket  : Bucket  : Bucket  : Bucket  : Bucket  : Bucket : Bucket  : Bucket  : P(  |  ) P(  |  )*P(  ),  =“yes” P(  | , ) P(  | ,  ),  =“yes” P(  |  =“yes”,  =“yes”) =  X\ {  } (P(  |  )* P(  |  )* P(  | , )* P(  | ,  )* P(  )*P( |  )*P(  |  )*P(  )) P( |  ) P(  |  )*P(  ) H()H() H()H() H(,)H(,) H  ( ,,  ) H ( , ,  ) H()H() H(,)H(,) P(  |  =“yes”,  =“yes”) H n (u)=  xn П j i=1 C i (x n,u si ) *k P (e) = 1- k, where k is a normalizing constant       

15 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Sensitivity Analysis Sensitivity analysis assesses how much the posterior probability of some event of interest changes with respect to the value of some parameter in the model We assume that the event of interest is the value of a target variable. The parameter is either a conditional probability or an unconditional prior probability If the sensitivity of the target variable having a particular value is low, then the analyst can be confident in the results, even if the analyst is not very confident in the precise value of the parameter If the sensitivity of the target variable to a parameter is very high, it is necessary to inform the analyst of the need to qualify the conclusion reached or to expend more resources to become more confident in the exact value of the parameter

16 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Example: Case Study #4 Computing Sensitivity This is the output of the Sensitivity Analysis program on a situation-specific scenario for Case Study #4. In the context of the information already acquired, i.e., travel to dangerous places, large transfers of money, etc., the parameter that links financial irregularities to being a suspect is much more important for assessing the belief in Ramazi being a terrorist than the parameter that links dangerous travel to being a suspect. The analyst may want to concentrate on assessing the first parameter precisely.

17 UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Sensitivity Analysis: Formal Definition Let the evidence be a set of findings: Let t be a parameter in the situation-specific scenario Then, [Castillo et al., 1997; Jensen, 2000] α and β can be determined by computing P(e) for two values of t More generally, if t is a set of parameters, then P(e)(t) is a linear function in each parameter in t, i.e., it is a multi-linear function of t Recall that Then, We can therefore compute the sensitivity of a target variable V to a parameter t by repeating the same computation with two values for the evidence set, viz. e and


Download ppt "UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Conflicts in Bayesian Networks January 23, 2007 Marco Valtorta"

Similar presentations


Ads by Google