Bayesian inference J. Daunizeau

Slides:



Advertisements
Similar presentations
J. Daunizeau Institute of Empirical Research in Economics, Zurich, Switzerland Brain and Spine Institute, Paris, France Bayesian inference.
Advertisements

Bayesian inference Lee Harrison York Neuroimaging Centre 01 / 05 / 2009.
Bayesian inference Jean Daunizeau Wellcome Trust Centre for Neuroimaging 16 / 05 / 2008.
Hierarchical Models and
Wellcome Trust Centre for Neuroimaging
Bayesian models for fMRI data
Bayesian Inference Chris Mathys Wellcome Trust Centre for Neuroimaging UCL SPM Course London, May 12, 2014 Thanks to Jean Daunizeau and Jérémie Mattout.
MEG/EEG Inverse problem and solutions In a Bayesian Framework EEG/MEG SPM course, Bruxelles, 2011 Jérémie Mattout Lyon Neuroscience Research Centre ? ?
Bayes for beginners Methods for dummies 27 February 2013 Claire Berna
Classical inference and design efficiency Zurich SPM Course 2014
Bayesian models for fMRI data
07/01/15 MfD 2014 Xin You Tai & Misun Kim
Bayesian models for fMRI data Methods & models for fMRI data analysis 06 May 2009 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK Institute of Empirical Research in Economics, Zurich, Switzerland Bayesian inference.
Machine Learning CMPT 726 Simon Fraser University
General Linear Model & Classical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM M/EEGCourse London, May.
Overview for Dummies Outline Getting started with an experiment Getting started with an experiment Things you need to know for scanning Things you need.
Methods for Dummies 2009 Bayes for Beginners Georgina Torbet & Raphael Kaplan.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Corinne Introduction/Overview & Examples (behavioral) Giorgia functional Brain Imaging Examples, Fixed Effects Analysis vs. Random Effects Analysis Models.
Bayesian Inference and Posterior Probability Maps Guillaume Flandin Wellcome Department of Imaging Neuroscience, University College London, UK SPM Course,
Practical Statistics for Particle Physicists Lecture 3 Harrison B. Prosper Florida State University European School of High-Energy Physics Anjou, France.
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK UZH – Foundations of Human Social Behaviour, Zurich, Switzerland Dynamic Causal Modelling:
Randomized Algorithms for Bayesian Hierarchical Clustering
J. Daunizeau ICM, Paris, France ETH, Zurich, Switzerland Dynamic Causal Modelling of fMRI timeseries.
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
Bayesian models for fMRI data Methods & models for fMRI data analysis November 2011 With many thanks for slides & images to: FIL Methods group, particularly.
Lecture 2: Statistical learning primer for biologists
EEG/MEG source reconstruction
Bayesian Methods Will Penny and Guillaume Flandin Wellcome Department of Imaging Neuroscience, University College London, UK SPM Course, London, May 12.
Bayesian inference Lee Harrison York Neuroimaging Centre 23 / 10 / 2009.
Mixture Models with Adaptive Spatial Priors Will Penny Karl Friston Acknowledgments: Stefan Kiebel and John Ashburner The Wellcome Department of Imaging.
Bayesian Inference in fMRI Will Penny Bayesian Approaches in Neuroscience Karolinska Institutet, Stockholm February 2016.
Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course Zurich, February 2008 Bayesian Inference.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Bayes for Beginners Anne-Catherine Huys M. Berk Mirza Methods for Dummies 20 th January 2016.
Bayesian Inference in SPM2 Will Penny K. Friston, J. Ashburner, J.-B. Poline, R. Henson, S. Kiebel, D. Glaser Wellcome Department of Imaging Neuroscience,
General Linear Model & Classical Inference London, SPM-M/EEG course May 2016 Sven Bestmann, Sobell Department, Institute of Neurology, UCL
Bayesian Model Selection and Averaging SPM for MEG/EEG course Peter Zeidman 17 th May 2016, 16:15-17:00.
J. Daunizeau ICM, Paris, France TNU, Zurich, Switzerland
Group Analyses Guillaume Flandin SPM Course London, October 2016
Dynamic Causal Modeling of Endogenous Fluctuations
Variational Bayesian Inference for fMRI time series
General Linear Model & Classical Inference
The general linear model and Statistical Parametric Mapping
Hypothesis Testing – Introduction
Wellcome Trust Centre for Neuroimaging
Neuroscience Research Institute University of Manchester
Dynamic Causal Model for evoked responses in M/EEG Rosalyn Moran.
Contrasts & Statistical Inference
'Linear Hierarchical Models'
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Linear Hierarchical Modelling
Statistical Parametric Mapping
The general linear model and Statistical Parametric Mapping
SPM2: Modelling and Inference
Bayesian Methods in Brain Imaging
Expectation-Maximization & Belief Propagation
Hierarchical Models and
M/EEG Statistical Analysis & Source Localization
Contrasts & Statistical Inference
Bayesian Inference in SPM2
Wellcome Centre for Neuroimaging, UCL, UK.
Mixture Models with Adaptive Spatial Priors
Will Penny Wellcome Trust Centre for Neuroimaging,
Bayesian Model Selection and Averaging
Contrasts & Statistical Inference
Presentation transcript:

Bayesian inference J. Daunizeau Institute of Empirical Research in Economics, Zurich, Switzerland Brain and Spine Institute, Paris, France

Overview of the talk 1 Probabilistic modelling and representation of uncertainty 1.1 Bayesian paradigm 1.2 Hierarchical models 1.3 Frequentist versus Bayesian inference 2 Numerical Bayesian inference methods 2.1 Sampling methods 2.2 Variational methods (ReML, EM, VB) 3 SPM applications 3.1 aMRI segmentation 3.2 Decoding of brain images 3.3 Model-based fMRI analysis (with spatial priors) 3.4 Dynamic causal modelling

Overview of the talk 1 Probabilistic modelling and representation of uncertainty 1.1 Bayesian paradigm 1.2 Hierarchical models 1.3 Frequentist versus Bayesian inference 2 Numerical Bayesian inference methods 2.1 Sampling methods 2.2 Variational methods (ReML, EM, VB) 3 SPM applications 3.1 aMRI segmentation 3.2 Decoding of brain images 3.3 Model-based fMRI analysis (with spatial priors) 3.4 Dynamic causal modelling

Bayesian paradigm probability theory: basics Degree of plausibility desiderata: - should be represented using real numbers (D1) - should conform with intuition (D2) - should be consistent (D3) a=2 • normalization: a=2 b=5 Bayesian theory of probabilities furnishes: - a language allowing to represent information and uncertainty. - calculus tools able to manipulate these notions in a coherent manner. It is an extension to formal logic, which works with propositions which are either true or false. Theory of probabilities rather works with propositions which are more or less true or more or less false. In other words, these propositions are characterized by their degree of plausibility. D2: e.g., a higher degree of plausibility should correspond to a higher number. D3: i.e. should not contain contradictions. E.g., when a result can be derived in different ways, all calculus should yield the same result. Finally, equal states of belief should correspond to equal degrees of plausibility. - plausibility can be measured on probability scale (sum to 1) • marginalization: • conditioning : (Bayes rule)

Bayesian paradigm deriving the likelihood function - Model of data with unknown parameters: e.g., GLM: - But data is noisy: - Assume noise/residuals is ‘small’: f Bayesian theory of probabilities furnishes: - a language allowing to represent information and uncertainty. - calculus tools able to manipulate these notions in a coherent manner. It is an extension to formal logic, which works with propositions which are either true or false. Theory of probabilities rather works with propositions which are more or less true or more or less false. In other words, these propositions are characterized by their degree of plausibility. D2: e.g., a higher degree of plausibility should correspond to a higher number. D3: i.e. should not contain contradictions. E.g., when a result can be derived in different ways, all calculus should yield the same result. Finally, equal states of belief should correspond to equal degrees of plausibility. - plausibility can be measured on probability scale (sum to 1) → Distribution of data, given fixed parameters:

Bayesian paradigm likelihood, priors and the model evidence Bayes rule: generative model m EXAMPLE: -Y : EEG electric potential measured on the scalp -theta : connection strength between two nodes of a network

Bayesian paradigm forward and inverse problems forward problem likelihood inverse problem posterior distribution 7

Bayesian paradigm model comparison Principle of parsimony : « plurality should not be assumed without necessity » y = f(x) x “Occam’s razor” : model evidence p(y|m) space of all data sets Model evidence: y=f(x)

Hierarchical models principle ••• hierarchy causality

Hierarchical models directed acyclic graphs (DAGs)

Hierarchical models univariate linear hierarchical model ••• prior densities posterior densities

Frequentist versus Bayesian inference a (quick) note on hypothesis testing if then reject H0 • estimate parameters (obtain test stat.) • define the null, e.g.: • apply decision rule, i.e.: classical SPM if then accept H0 • invert model (obtain posterior pdf) • define the null, e.g.: • apply decision rule, i.e.: Bayesian PPM

Frequentist versus Bayesian inference what about bilateral tests? • define the null and the alternative hypothesis in terms of priors, e.g.: space of all datasets • apply decision rule, i.e.: if then reject H0 Savage-Dickey ratios (nested models, i.i.d. priors):

Overview of the talk 1 Probabilistic modelling and representation of uncertainty 1.1 Bayesian paradigm 1.2 Hierarchical models 1.3 Frequentist versus Bayesian inference 2 Numerical Bayesian inference methods 2.1 Sampling methods 2.2 Variational methods (ReML, EM, VB) 3 SPM applications 3.1 aMRI segmentation 3.2 Decoding of brain images 3.3 Model-based fMRI analysis (with spatial priors) 3.4 Dynamic causal modelling

MCMC example: Gibbs sampling Sampling methods MCMC example: Gibbs sampling

Variational methods VB / EM / ReML → VB : maximize the free energy F(q) w.r.t. the “variational” posterior q(θ) under some (e.g., mean field, Laplace) approximation

Overview of the talk 1 Probabilistic modelling and representation of uncertainty 1.1 Bayesian paradigm 1.2 Hierarchical models 1.3 Frequentist versus Bayesian inference 2 Numerical Bayesian inference methods 2.1 Sampling methods 2.2 Variational methods (ReML, EM, VB) 3 SPM applications 3.1 aMRI segmentation 3.2 Decoding of brain images 3.3 Model-based fMRI analysis (with spatial priors) 3.4 Dynamic causal modelling

posterior probability segmentation and normalisation posterior probability maps (PPMs) dynamic causal modelling multivariate decoding realignment smoothing general linear model statistical inference Gaussian field theory normalisation p <0.05 template

aMRI segmentation mixture of Gaussians (MoG) model … class variances class means ith voxel value label frequencies grey matter CSF white matter

Decoding of brain images recognizing brain states from fMRI + fixation cross >> pace response log-evidence of X-Y sparse mappings: effect of lateralization log-evidence of X-Y bilateral mappings: effect of spatial deployment

fMRI time series analysis spatial priors and model comparison PPM: regions best explained by short-term memory model short-term memory design matrix (X) fMRI time series GLM coeff prior variance of GLM coeff of data noise AR coeff (correlated noise) PPM: regions best explained by long-term memory model long-term memory design matrix (X)

Dynamic Causal Modelling network structure identification attention attention attention attention PPC PPC PPC PPC stim V1 V5 stim V1 V5 stim V1 V5 stim V1 V5 V1 V5 stim PPC attention 1.25 0.13 0.46 0.39 0.26 0.10 estimated effective synaptic strengths for best model (m4) models marginal likelihood m1 m2 m3 m4 15 10 5

DCMs and DAGs a note on causality 1 2 1 2 3 1 2 1 2 3 time 3 3 u

Dynamic Causal Modelling model comparison for group studies differences in log- model evidences m2 subjects fixed effect assume all subjects correspond to the same model random effect assume different subjects might correspond to different models

I thank you for your attention.

A note on statistical significance lessons from the Neyman-Pearson lemma Neyman-Pearson lemma: the likelihood ratio (or Bayes factor) test is the most powerful test of size to test the null. what is the threshold u, above which the Bayes factor test yields a error I rate of 5%? ROC analysis MVB (Bayes factor) u=1.09, power=56% 1 - error II rate CCA (F-statistics) F=2.20, power=20% error I rate