Graphical models for combining multiple sources of information in observational studies Nicky Best Sylvia Richardson Chris Jackson Virgilio Gomez Sara.

Slides:



Advertisements
Similar presentations
Sources and effects of bias in investigating links between adverse health outcomes and environmental hazards Frank Dunstan University of Wales College.
Advertisements

Gated Graphs and Causal Inference
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
A Tutorial on Learning with Bayesian Networks
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
Comments on Hierarchical models, and the need for Bayes Peter Green, University of Bristol, UK IWSM, Chania, July 2002.
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
BACKGROUND Benzene is a known carcinogen. Occupational exposure to benzene is an established risk factor for leukaemia. Less is known about the effects.
Dynamic Bayesian Networks (DBNs)
Nicky Best, Chris Jackson, Sylvia Richardson Department of Epidemiology and Public Health Imperial College, London Studying.
IMPORTANCE SAMPLING ALGORITHM FOR BAYESIAN NETWORKS
Chapter 4: Linear Models for Classification
1 Slides for the book: Probabilistic Robotics Authors: Sebastian Thrun Wolfram Burgard Dieter Fox Publisher: MIT Press, Web site for the book & more.
Chance, bias and confounding
Nicky Best and Chris Jackson With Sylvia Richardson Department of Epidemiology and Public Health Imperial College, London
Data Mining Cardiovascular Bayesian Networks Charles Twardy †, Ann Nicholson †, Kevin Korb †, John McNeil ‡ (Danny Liew ‡, Sophie Rogers ‡, Lucas Hope.
Bayesian network inference
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Part III: Inference Topic 6 Sampling and Sampling Distributions
. Approximate Inference Slides by Nir Friedman. When can we hope to approximate? Two situations: u Highly stochastic distributions “Far” evidence is discarded.
Bayesian Filtering for Location Estimation D. Fox, J. Hightower, L. Liao, D. Schulz, and G. Borriello Presented by: Honggang Zhang.
Causal Models, Learning Algorithms and their Application to Performance Modeling Jan Lemeire Parallel Systems lab November 15 th 2006.
Sampling Theory Determining the distribution of Sample statistics.
Statistical Methods for Missing Data Roberta Harnett MAR 550 October 30, 2007.
Learning In Bayesian Networks. Learning Problem Set of random variables X = {W, X, Y, Z, …} Training set D = { x 1, x 2, …, x N }  Each observation specifies.
Aspects of Bayesian Inference and Statistical Disclosure Control in Python Duncan Smith Confidentiality and Privacy Group CCSR University of Manchester.
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
Bayesian Decision Theory Making Decisions Under uncertainty 1.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Hierarchical models for combining multiple data sources measured at individual and small area levels Chris Jackson With Nicky Best and Sylvia Richardson.
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Graphical models for combining multiple data sources
Chris Jackson With Nicky Best and Sylvia Richardson Department of Epidemiology and Public Health Imperial College, London
Nuoo-Ting (Jassy) Molitor 1 Chris Jackson 2 With Nicky Best, Sylvia Richardson 1 1 Department of Epidemiology and Public Health Imperial College, London.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Visibility Graph. Voronoi Diagram Control is easy: stay equidistant away from closest obstacles.
Areej Jouhar & Hafsa El-Zain Biostatistics BIOS 101 Foundation year.
Nuoo-Ting (Jassy) Molitor 1 Chris Jackson 2 With Nicky Best, Sylvia Richardson 1 1 Department of Epidemiology and Public Health Imperial College, London.
Chapter 2 Nature of the evidence. Chapter overview Introduction What is epidemiology? Measuring physical activity and fitness in population studies Laboratory-based.
BACKGROUND Benzene is a known carcinogen. Occupational exposure to benzene is an established risk factor for leukaemia. Less is known about the effects.
Lecture 13: Linkage Analysis VI Date: 10/08/02  Complex models  Pedigrees  Elston-Stewart Algorithm  Lander-Green Algorithm.
INTERVENTIONS AND INFERENCE / REASONING. Causal models  Recall from yesterday:  Represent relevance using graphs  Causal relevance ⇒ DAGs  Quantitative.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Learning In Bayesian Networks. General Learning Problem Set of random variables X = {X 1, X 2, X 3, X 4, …} Training set D = { X (1), X (2), …, X (N)
CHAPTER 5 Probability Theory (continued) Introduction to Bayesian Networks.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Methods and software for editing and imputation: recent advancements at Istat M. Di Zio, U. Guarnera, O. Luzi, A. Manzari ISTAT – Italian Statistical Institute.
Random Variables Numerical Quantities whose values are determine by the outcome of a random experiment.
Review of statistical modeling and probability theory Alan Moses ML4bio.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Classification COMP Seminar BCB 713 Module Spring 2011.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Probabilistic Robotics Probability Theory Basics Error Propagation Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics.
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
Prediction and Missing Data. Summarising Distributions ● Models are often large and complex ● Often only interested in some parameters – e.g. not so interested.
Review of Probability.
Epidemiology 503 Confounding.
Learning Bayesian Network Models from Data
More about Posterior Distributions
CSE-490DF Robotics Capstone
Pattern Recognition and Image Analysis
CSCI 5822 Probabilistic Models of Human and Machine Learning
Markov Networks.
Inferential statistics Study a sample Conclude about the population Two processes: Estimation (Point or Interval) Hypothesis testing.
CS639: Data Management for Data Science
Presentation transcript:

Graphical models for combining multiple sources of information in observational studies Nicky Best Sylvia Richardson Chris Jackson Virgilio Gomez Sara Geneletti ESRC National Centre for Research Methods – BIAS node

Outline Overview of graphical modelling Case study 1: Water disinfection byproducts and adverse birth outcomes –Modelling multiple sources of bias in observational studies Bayesian computation and software Case study 2: Socioeconomic factors and heart disease (Chris Jackson) –Combining individual and aggregate level data –Application to Census, Health Survey for England, HES

Graphical modelling Modelling Inference Mathematics Algorithms

1. Mathematics Modelling Inference Mathematics Algorithms Key idea: conditional independence X and W are conditionally independent given Z if, knowing Z, discovering W tells you nothing more about X P(X | W, Z) = P(X | Z)

Example: Mendelian inheritance Y, Z = genotype of parents W, X = genotypes of 2 children If we know the genotypes of the parents, then the children’s genotypes are conditionally independent P(X | W, Y, Z) = P(X | Y, Z) Y W Z X

Joint distributions and graphical models Graphical models can be used to: represent structure of a joint probability distribution….. …..by encoding conditional independencies Factorization thm: Jt distribution P(V) =  P(v | parents[v]) Y W Z X P(X|Y, Z) P(W|Y, Z) P(Z)P(Y) P(W,X,Y,Z) = P(W|Y,Z) P(X|Y,Z) P(Y) P(Z)

Where does the graph come from? Genetics –pedigree (family tree) Physical, biological, social systems –supposed causal effects (e.g. regression models)

Conditional independence provides basis for splitting large system into smaller components Y W Z X AB D C

Y W Z W D C Y Z X Y AB

2. Modelling Modelling Inference Mathematics Algorithms

Building complex models Key idea understand complex system through global model built from small pieces –comprehensible –each with only a few variables –modular

Example: Case study 1 Epidemiological study of low birth weight and mothers’ exposure to water disinfection byproducts Background – Chlorine added to tap water supply for disinfection –Reacts with natural organic matter in water to form unwanted byproducts (including trihalomethanes, THMs) –Some evidence of adverse health effects (cancer, birth defects) associated with exposure to high levels of THM –SAHSU are carrying out study in Great Britain using routine data, to investigate risk of low birth weight associated with exposure to different THM levels

Data sources National postcoded births register Routinely monitored THM concentrations in tap water samples for each water supply zone within 14 different water company regions Census data – area level socioeconomic factors Millenium cohort study (MCS) – individual level outcomes and confounder data on sample of mothers Literature relating to factors affecting personal exposure (uptake factors, water consumption, etc.)

Model for combining data sources  [c]  [T] y ik 22 y im c ik ii c im THM ik [mother] THM zt [true] THM ztj [raw] THM im [mother] 

Regression sub-model (MCS)  [c]  [T] y ik 22 y im c ik ii c im THM ik [mother] THM zt [true] THM ztj [raw] THM im [mother]  Regression model for MCS data relating risk of low birth weight (y im ) to mother’s THM exposure and other confounders (c im )

Regression sub-model (MCS)  [c]  [T] y im c im THM im [mother] Regression model for MCS data relating risk of low birth weight (y im ) to mother’s THM exposure and other confounders (c im ) Logistic regression y im ~ Bernoulli(p im ) logit p im = b [c] c im + b [T] THM im i indexes small area m indexes mother [mother] c ik = potential confounders, e.g. deprivation, smoking, ethnicity

Regression sub-model (national data)  [c]  [T] y ik 22 y im c ik ii c im THM ik [mother] THM zt [true] THM ztj [raw] THM im [mother]  Regression model for national data relating risk of low birth weight (y ik ) to mother’s THM exposure and other confounders (c ik )

Regression sub-model (national data)  [c]  [T] y ik c ik THM ik [mother] Regression model for national data relating risk of low birth weight (y ik ) to mother’s THM exposure and other confounders (c ik ) Logistic regression y ik ~ Bernoulli(p ik ) logit p ik = b [c] c ik + b [T] THM ik i indexes small area k indexes mother [mother]

Missing confounders sub-model  [c]  [T] y ik 22 y im c ik ii c im THM ik [mother] THM zt [true] THM ztj [raw] THM im [mother]  Missing data model to estimate confounders (c ik ) for mothers in national data, using information on within area distribution of confounders in MCS

Missing confounders sub-model c ik ii c im Missing data model to estimate confounders (c ik ) for mothers in national data, using information on within area distribution of confounders in MCS c im ~ Bernoulli(  i ) (MCS mothers) c ik ~ Bernoulli(  i ) (Predictions for mothers in national data)

THM measurement error sub-model  [c]  [T] y ik 22 y im c ik ii c im THM ik [mother] THM zt [true] THM ztj [raw] THM im [mother]  Model to estimate true tap water THM concentration from raw data

THM measurement error sub-model 22 THM zt [true] THM ztj [raw] Model to estimate true tap water THM concentration from raw data THM ztj ~ Normal(THM zt,  2 ) z = water zone; t = season; j = sample (Actual model used was a more complex mixture of Normal distributions) [raw][true]

THM personal exposure sub-model  [c]  [T] y ik 22 y im c ik ii c im THM ik [mother] THM zt [true] THM ztj [raw] THM im [mother]  Model to predict personal exposure using estimated tap water THM level and literature on distribution of factors affecting individual uptake of THM

THM personal exposure sub-model THM ik [mother] THM zt [true] THM im [mother]  Model to predict personal exposure using estimated tap water THM level and literature on distribution of factors affecting individual uptake of THM THM = ∑ k THM zt x quantity (  1k ) x uptake factor (  2k ) where k indexes different water use activities, e.g. drinking, showering, bathing [mother][true]

3. Inference Modelling Inference Mathematics Algorithms

Bayesian

… or non Bayesian

Graphical approach to building complex models lends itself naturally to Bayesian inferential process Graph defines joint probability distribution on all the ‘nodes’ in the model Recall: Joint distribution P(V) =  P(v | parents[v]) Condition on parts of graph that are observed (data) Calculate posterior probabilities of remaining nodes using Bayes theorem Automatically propagates all sources of uncertainty Bayesian Full Probability Modelling

 [c]  [T] y ik 22 y im c ik ii c im THM ik [mother] THM zt [true] THM ztj [raw] THM im [mother]  Data Unknowns

4. Algorithms Modelling Inference Mathematics Algorithms MCMC algorithms are able to exploit graphical structure for efficient inference Bayesian graphical models implemented in WinBUGS