Digital Statisticians INST 4200 David J Stucki Spring 2015.

Slides:



Advertisements
Similar presentations
Decision Making Under Risk Continued: Bayes’Theorem and Posterior Probabilities MGS Chapter 8 Slides 8c.
Advertisements

1 Bayesian Networks Slides from multiple sources: Weng-Keen Wong, School of Electrical Engineering and Computer Science, Oregon State University.
Uncertainty Everyday reasoning and decision making is based on uncertain evidence and inferences. Classical logic only allows conclusions to be strictly.
FT228/4 Knowledge Based Decision Support Systems
1 Knowledge Engineering for Bayesian Networks. 2 Probability theory for representing uncertainty l Assigns a numerical degree of belief between 0 and.
Homework 3: Naive Bayes Classification
Foundations of Artificial Intelligence 1 Bayes’ Rule - Example  Medical Diagnosis  suppose we know from statistical data that flu causes fever in 80%
1 A Tutorial on Bayesian Networks Weng-Keen Wong School of Electrical Engineering and Computer Science Oregon State University.
Intro to Bayesian Learning Exercise Solutions Ata Kaban The University of Birmingham 2005.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. 6.1 Chapter Six Probability.
Bayesian network inference
 2004 University of Pittsburgh Bayesian Biosurveillance Using Multiple Data Streams Weng-Keen Wong, Greg Cooper, Denver Dash *, John Levander, John Dowling,
Solved the Maze? Start at phil’s house. At first, you can only make right turns through the maze. Each time you cross the red zigzag sign (under Carl’s.
 2004 University of Pittsburgh Bayesian Biosurveillance Using Multiple Data Streams Greg Cooper, Weng-Keen Wong, Denver Dash*, John Levander, John Dowling,
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 12 Jim Martin.
Bayesian Belief Networks
AI - Week 24 Uncertain Reasoning (quick mention) then REVISION Lee McCluskey, room 2/07
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
Weng-Keen Wong, Oregon State University © Bayesian Networks: A Tutorial Weng-Keen Wong School of Electrical Engineering and Computer Science Oregon.
Inductive Reasoning Bayes Rule. Urn problem (1) A B A die throw determines from which urn to select balls. For outcomes 1,2, and 3, balls are picked from.
CSE (c) S. Tanimoto, 2008 Bayes Nets 1 Probabilistic Reasoning With Bayes’ Rule Outline: Motivation Generalizing Modus Ponens Bayes’ Rule Applying.
Conclusions On our large scale anthrax attack simulations, being able to infer the work zip appears to improve detection time over just using the home.
Population-Wide Anomaly Detection Weng-Keen Wong 1, Gregory Cooper 2, Denver Dash 3, John Levander 2, John Dowling 2, Bill Hogan 2, Michael Wagner 2 1.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
1 A Tutorial on Bayesian Networks Modified by Paul Anderson from slides by Weng-Keen Wong School of Electrical Engineering and Computer Science Oregon.
Bayesian Networks Textbook: Probabilistic Reasoning, Sections 1-2, pp
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
CISC 4631 Data Mining Lecture 06: Bayes Theorem Theses slides are based on the slides by Tan, Steinbach and Kumar (textbook authors) Eamonn Koegh (UC Riverside)
Soft Computing Lecture 17 Introduction to probabilistic reasoning. Bayesian nets. Markov models.
Dr. Gary Blau, Sean HanMonday, Aug 13, 2007 Statistical Design of Experiments SECTION I Probability Theory Review.
Lecture 5a: Bayes’ Rule Class web site: DEA in Bioinformatics: Statistics Module Box 1Box 2Box 3.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Visibility Graph. Voronoi Diagram Control is easy: stay equidistant away from closest obstacles.
“PREDICTIVE MODELING” CoSBBI, July Jennifer Hu.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
BINOMIALDISTRIBUTION AND ITS APPLICATION. Binomial Distribution  The binomial probability density function –f(x) = n C x p x q n-x for x=0,1,2,3…,n for.
1 Monte Carlo Artificial Intelligence: Bayesian Networks.
CSE PR 1 Reasoning - Rule-based and Probabilistic Representing relations with predicate logic Limitations of predicate logic Representing relations.
Bayesian Classification. Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership probabilities.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Week 21 Conditional Probability Idea – have performed a chance experiment but don’t know the outcome (ω), but have some partial information (event A) about.
Bayes’ Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Review: Probability Random variables, events Axioms of probability Atomic events Joint and marginal probability distributions Conditional probability distributions.
Inference Algorithms for Bayes Networks
1 Bayesian Networks: A Tutorial. 2 Introduction Suppose you are trying to determine if a patient has tuberculosis. You observe the following symptoms:
Quiz 3: Mean: 9.2 Median: 9.75 Go over problem 1.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Bayes network inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y 
Conditional Probability, Bayes’ Theorem, and Belief Networks CISC 2315 Discrete Structures Spring2010 Professor William G. Tanner, Jr.
Bayesian networks Chapter 14 Slide Set 2. Constructing Bayesian networks 1. Choose an ordering of variables X 1, …,X n 2. For i = 1 to n –add X i to the.
HL2 Math - Santowski Lesson 93 – Bayes’ Theorem. Bayes’ Theorem  Main theorem: Suppose we know We would like to use this information to find if possible.
Weng-Keen Wong, Oregon State University © Bayesian Networks: A Tutorial Weng-Keen Wong School of Electrical Engineering and Computer Science Oregon.
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Bayes’ Theorem Suppose we have estimated prior probabilities for events we are concerned with, and then obtain new information. We would like to a sound.
Biostatistics Class 2 Probability 2/1/2000.
Bayes Rule and Bayes Classifiers
Bayesian Networks: A Tutorial
Bayes for Beginners Stephanie Azzopardi & Hrvoje Stojic
Conditional Probability, Bayes’ Theorem, and Belief Networks
Bayesian Networks Probability In AI.
Learn to let go. That is the key to happiness. ~Jack Kornfield
Wellcome Trust Centre for Neuroimaging
Bayes for Beginners Luca Chech and Jolanda Malamud
CS 594: Empirical Methods in HCC Introduction to Bayesian Analysis
Presentation transcript:

Digital Statisticians INST 4200 David J Stucki Spring 2015

Weng-Keen Wong, Oregon State University © Introduction Suppose you are trying to determine if a patient has inhalational anthrax. You observe the following symptoms: The patient has a cough The patient has a fever The patient has difficulty breathing

Weng-Keen Wong, Oregon State University © Introduction You would like to determine how likely the patient is infected with inhalational anthrax given that the patient has a cough, a fever, and difficulty breathing We are not 100% certain that the patient has anthrax because of these symptoms. We are dealing with uncertainty!

Weng-Keen Wong, Oregon State University © Introduction Now suppose you order an x-ray and observe that the patient has a wide mediastinum. Your belief that that the patient is infected with inhalational anthrax is now much higher.

Weng-Keen Wong, Oregon State University © Introduction In the previous slides, what you observed affected your belief that the patient is infected with anthrax This is called reasoning with uncertainty Wouldn’t it be nice if we had some methodology for reasoning with uncertainty? Why in fact, we do…

Weng-Keen Wong, Oregon State University © Probabilities The sum of the red and blue areas is 1 P(A = false) P(A = true) We will write P(A = true) to mean the probability that A = true. What is probability? It is the relative frequency with which an outcome would be obtained if the process were repeated a large number of times under similar conditions * * Ahem…there’s also the Bayesian definition which says probability is your degree of belief in an outcome

Introduction - Bayes’ Theorem 7 Test A (test to screen for disease X)  Prevalence of disease X was 0.3%  Sensitivity (true positive) of the test was 50%  False positive rate was 3%. What is the probability that someone who tests positive actually has disease X?  Doctors’ answers ranged from 1% to 99% (with ~half of them estimating the probability as 50% or 47%) Gerd Gigerenzer, Adrian Edwards “Simple tools for understanding risks: from innumeracy to insight” BMJ VOLUME 327 (2003) Francois Ayello, Andrea Sanchez, Vinod Khare, DNV GL ©2015

Test Terminology

Introduction - Bayes’ Theorem 9 Test A (test to screen for disease X)  Prevalence of disease X was 0.3%  Sensitivity (true positive) of the test was 50%  False positive rate was 3%. What is the probability that someone who tests positive actually has disease X?  Doctors’ answers ranged from 1% to 99% (with ~half of them estimating the probability as 50% or 47%)  The correct answer is ~5%! Gerd Gigerenzer, Adrian Edwards “Simple tools for understanding risks: from innumeracy to insight” BMJ VOLUME 327 (2003) Francois Ayello, Andrea Sanchez, Vinod Khare, DNV GL ©2015

Let’s do the math… What is the probability that someone who tests positive actually has disease X? Francois Ayello, Andrea Sanchez, Vinod Khare, DNV GL ©2015

Weng-Keen Wong, Oregon State University © Conditional Probability P(A | B) = Out of all the outcomes in which B is true, how many also have A equal to true Read this as: “Probability of A conditioned on B” or “Probability of A given B” P(F) P(H) H = “Have a headache” F = “Coming down with Flu” P(H) = 1/10 P(F) = 1/40 P(H | F) = 1/2 “Headaches are rare and flu is rarer, but if you’re coming down with flu there’s a chance you’ll have a headache.”

Bayes’ Theorem 12 where A and B are events. P(A) and P(B) are the probabilities of A and B independent of each other. P(A|B), a conditional probability, is the probability of A given that B is true. P(B|A), is the probability of B given that A is true.

Bayes’ Theorem 13 Belief Prior distribution Evidence (observed data) Posterior distribution The prior distribution is the probability value that the person has before observing data. Disease X Tests The posterior distribution is the probability value that has been revised by using additional information that is later obtained. Francois Ayello, Andrea Sanchez, Vinod Khare, DNV GL ©2015

Bayesian Networks Disease X Tests Belief Prior distribution Evidence (observed data) Posterior distribution BAYESIAN NETWORKS Francois Ayello, Andrea Sanchez, Vinod Khare, DNV GL ©2015

Bayesian Networks Disease X Tests Belief DISEASE Yes0.003 No0.997 DISEASEYesNo TEST Positive Negative Belief Prior distribution Evidence (observed data) Posterior distribution BAYESIAN NETWORKS Francois Ayello, Andrea Sanchez, Vinod Khare, DNV GL ©2015

So let’s calculate it out…

A Bayesian Network A Bayesian network is made up of: AP(A) false0.6 true0.4 A B CD ABP(B|A) false 0.01 falsetrue0.99 truefalse0.7 true 0.3 BCP(C|B) false 0.4 falsetrue0.6 truefalse0.9 true 0.1 BDP(D|B) false 0.02 falsetrue0.98 truefalse0.05 true A Directed Acyclic Graph 2. A set of tables for each node in the graph

Weng-Keen Wong, Oregon State University © A Directed Acyclic Graph A B CD Each node in the graph is a random variable A node X is a parent of another node Y if there is an arrow from node X to node Y eg. A is a parent of B Informally, an arrow from node X to node Y means X has a direct influence on Y

A Set of Tables for Each Node Each node X i has a conditional probability distribution P(X i | Parents(X i )) that quantifies the effect of the parents on the node The parameters are the probabilities in these conditional probability tables (CPTs) AP(A) false0.6 true0.4 ABP(B|A) false 0.01 falsetrue0.99 truefalse0.7 true 0.3 BCP(C|B) false 0.4 falsetrue0.6 truefalse0.9 true 0.1 BDP(D|B) false 0.02 falsetrue0.98 truefalse0.05 true 0.95 A B CD

Weng-Keen Wong, Oregon State University © Inference Using a Bayesian network to compute probabilities is called inference In general, inference involves queries of the form: P( X | E ) X = The query variable(s) E = The evidence variable(s)

Weng-Keen Wong, Oregon State University © Inference An example of a query would be: P( HasAnthrax| HasFever and HasCough) Note: Even though HasDifficultyBreathing and HasWideMediastinum are in the Bayesian network, they are not given values in the query (ie. they do not appear either as query variables or evidence variables) They are treated as unobserved variables HasAnthrax HasCoughHasFeverHasDifficultyBreathingHasWideMediastinum

Weng-Keen Wong, Oregon State University © The Bad News Exact inference is feasible in small to medium-sized networks Exact inference in large networks takes a very long time We resort to approximate inference techniques which are much faster and give pretty good results

Person Model (Initial Prototype) Anthrax Release Location of ReleaseTime Of Release Anthrax Infection Home Zip Respiratory from Anthrax Other ED Disease GenderAge Decile Respiratory CC From Other Respiratory CC Respiratory CC When Admitted ED Admit from Anthrax ED Admit from Other ED Admission Anthrax Infection Home Zip Respiratory from Anthrax Other ED Disease Gender Age Decile Respiratory CC From Other Respiratory CC Respiratory CC When Admitted ED Admit from Anthrax ED Admit from Other ED Admission … … Yesterdaynever False Female Unknown Male

Weng-Keen Wong, Oregon State University © Bayesian Networks In the opinion of many AI researchers, Bayesian networks are the most significant contribution in AI in the last 10 years They are used in many applications eg. spam filtering, speech recognition, robotics, diagnostic systems and even syndromic surveillance HasAnthrax HasCoughHasFeverHasDifficultyBreathingHasWideMediastinum