Estimating the Transfer Function from Neuronal Activity to BOLD Maria Joao Rosa SPM Homecoming 2008 Wellcome Trust Centre for Neuroimaging.

Slides:



Advertisements
Similar presentations
J. Daunizeau Institute of Empirical Research in Economics, Zurich, Switzerland Brain and Spine Institute, Paris, France Bayesian inference.
Advertisements

Bayesian inference Lee Harrison York Neuroimaging Centre 01 / 05 / 2009.
Bayesian inference Jean Daunizeau Wellcome Trust Centre for Neuroimaging 16 / 05 / 2008.
Hierarchical Models and
Bayesian models for fMRI data
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
CHAPTER 21 Inferential Statistical Analysis. Understanding probability The idea of probability is central to inferential statistics. It means the chance.
MEG/EEG Inverse problem and solutions In a Bayesian Framework EEG/MEG SPM course, Bruxelles, 2011 Jérémie Mattout Lyon Neuroscience Research Centre ? ?
Bayes for beginners Methods for dummies 27 February 2013 Claire Berna
Classical inference and design efficiency Zurich SPM Course 2014
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich With.
Bayesian models for fMRI data
07/01/15 MfD 2014 Xin You Tai & Misun Kim
Bayesian models for fMRI data Methods & models for fMRI data analysis 06 May 2009 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK Institute of Empirical Research in Economics, Zurich, Switzerland Bayesian inference.
Elementary hypothesis testing Purpose of hypothesis testing Type of hypotheses Type of errors Critical regions Significant levels Hypothesis vs intervals.
General Linear Model & Classical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM M/EEGCourse London, May.
Random Field Theory Will Penny SPM short course, London, May 2005 Will Penny SPM short course, London, May 2005 David Carmichael MfD 2006 David Carmichael.
Graziella Quattrocchi & Louise Marshall Methods for Dummies 2014
Dynamic Causal Modelling (DCM): Theory Demis Hassabis & Hanneke den Ouden Thanks to Klaas Enno Stephan Functional Imaging Lab Wellcome Dept. of Imaging.
Bayes for Beginners Presenters: Shuman ji & Nick Todd.
METHODSDUMMIES BAYES FOR BEGINNERS. Any given Monday at pm “I’m sure this makes sense, but you lost me about here…”
Methods for Dummies 2009 Bayes for Beginners Georgina Torbet & Raphael Kaplan.
Dr. Gary Blau, Sean HanMonday, Aug 13, 2007 Statistical Design of Experiments SECTION I Probability Theory Review.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
General Linear Model & Classical Inference London, SPM-M/EEG course May 2014 C. Phillips, Cyclotron Research Centre, ULg, Belgium
SPM Course Zurich, February 2015 Group Analyses Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London With many thanks to.
Bayes for Beginners Reverend Thomas Bayes ( ) Velia Cardin Marta Garrido.
Corinne Introduction/Overview & Examples (behavioral) Giorgia functional Brain Imaging Examples, Fixed Effects Analysis vs. Random Effects Analysis Models.
Hypothesis Testing A procedure for determining which of two (or more) mutually exclusive statements is more likely true We classify hypothesis tests in.
G. Cowan Lectures on Statistical Data Analysis Lecture 1 page 1 Lectures on Statistical Data Analysis London Postgraduate Lectures on Particle Physics;
Bayesian Inference and Posterior Probability Maps Guillaume Flandin Wellcome Department of Imaging Neuroscience, University College London, UK SPM Course,
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove.
Classical Inference on SPMs Justin Chumbley SPM Course Oct 23, 2008.
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
DCM – the theory. Bayseian inference DCM examples Choosing the best model Group analysis.
Contrasts & Statistical Inference
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
Bayesian models for fMRI data Methods & models for fMRI data analysis November 2011 With many thanks for slides & images to: FIL Methods group, particularly.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
Ch. 5 Bayesian Treatment of Neuroimaging Data Will Penny and Karl Friston Ch. 5 Bayesian Treatment of Neuroimaging Data Will Penny and Karl Friston 18.
Dynamic Causal Model for evoked responses in MEG/EEG Rosalyn Moran.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
Bayesian Methods Will Penny and Guillaume Flandin Wellcome Department of Imaging Neuroscience, University College London, UK SPM Course, London, May 12.
Bayesian inference Lee Harrison York Neuroimaging Centre 23 / 10 / 2009.
Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course Zurich, February 2008 Bayesian Inference.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Bayes for Beginners Anne-Catherine Huys M. Berk Mirza Methods for Dummies 20 th January 2016.
Bayesian Inference in SPM2 Will Penny K. Friston, J. Ashburner, J.-B. Poline, R. Henson, S. Kiebel, D. Glaser Wellcome Department of Imaging Neuroscience,
General Linear Model & Classical Inference Short course on SPM for MEG/EEG Wellcome Trust Centre for Neuroimaging University College London May 2010 C.
Bayesian Model Selection and Averaging SPM for MEG/EEG course Peter Zeidman 17 th May 2016, 16:15-17:00.
J. Daunizeau ICM, Paris, France TNU, Zurich, Switzerland
Group Analyses Guillaume Flandin SPM Course London, October 2016
Bayesian Inference Will Penny
Bayes for Beginners Stephanie Azzopardi & Hrvoje Stojic
Hypothesis Testing: Hypotheses
Statistical NLP: Lecture 4
Statistical Parametric Mapping
SPM2: Modelling and Inference
Wellcome Trust Centre for Neuroimaging
Hierarchical Models and
Bayes for Beginners Luca Chech and Jolanda Malamud
Bayesian inference J. Daunizeau
Bayesian Inference in SPM2
Wellcome Centre for Neuroimaging, UCL, UK.
Will Penny Wellcome Trust Centre for Neuroimaging,
Mathematical Foundations of BME Reza Shadmehr
Presentation transcript:

Estimating the Transfer Function from Neuronal Activity to BOLD Maria Joao Rosa SPM Homecoming 2008 Wellcome Trust Centre for Neuroimaging

Statistic formulations P(A): probability of event A occurring P(A|B): probability of A occurring given B occurred P(B|A): probability of B occurring given A occurred P(A,B): probability of A and B occurring simultaneously (joint probability of A and B) Joint probability of A and B P(A,B) = P(A|B)*P(B) = P(B|A)*P(A)  P(B|A) = P(A|B)*P(B)/P(A) Which is Bayes Rule Bayes’ Rule is very often referred to Bayes’ Theorem, but it is not really a theorem, and should more properly be referred to as Bayes’ Rule (Hacking, 2001).

Reverend Thomas Bayes (1702 – 1761) Reverend Thomas Bayes was a minister interested in probability and stated a form of his famous rule in the context of solving a somewhat complex problem involving billiard balls It was first stated by Bayes in his ‘Essay towards solving a problem in the doctrine of chances’, published in the Philosophical Transactions of the Royal Society of London in 1764.

Conditional probability P(A|B): conditional probability of A given B Q: When are we considering conditional probabilities? A: Almost always! Examples: Lottery chances Dice tossing

Conditional probability Examples (cont’): P(Brown eyes|Male): (P(A|B) with A := Brown eyes, B := Male) 1.What is the probability that a person has brown eyes, ignoring everyone who is not a male? 2.Ratio: (being a male with brown eyes)/(being a male) 3.Probability ratio: probability that a person is both male and has brown eyes to the probability that a person is male P(Male) = P(B) = 0.52 P(Brown eyes) = P(A) = 0.78 P(Male with brown eyes) = P(A,B) = 0.38 P(A|B) = P(B|A)*P(A)/P(B) = P(A,B)/P(B) = 0.38/0.52 = Flipping it around (Bayes idea): prob. of being a male if you have brown eyes You could also calculate now what’s the prob. of being a male if you have brown eyes P(B|A) = P(A|B)*P(B)/P(A) = 0.73*0.52/0.78 = …

Statistic terminology P(A) is called the marginal or prior probability of A (since it is the probability of A prior to having any information about B) Similarly: P(B): the marginal or prior probability of B P(A|B) is called the likelihood function for A given B. P(B|A): the posterior probability of B given A (since it depends on having information about A) Bayes Rule P(B|A) = P(A|B)*P(B)/P(A) “likelihood” function for B (for fixed A) “posterior” probability of B given A prior probabilities of B, A (“priors”)

It relates to the conditional density of a parameter (posterior probability) with its unconditional density (prior, since depends on information present before the experiment). The likelihood is the probability of the data given the parameter and represents the data now available. Bayes’ Theorem for a given parameter  p (  data) = p (data  ) p (  ) / p (data) 1/P (data) is basically a normalizing constant Posterior  likelihood x prior The prior is the probability of the parameter and represents what was thought before seeing the data. The posterior represents what is thought given both prior information and the data just seen.

Data and hypotheses… –We have a hypotheses H 0 (null), H 1 –We have data (Y) –We want to check if the model that we have (H 1 ) fits our data (accept H 1 / reject H 0 ) or not (H 0 ) Inferential statistics: what is the probability that we can reject H 0 and accept H 1 at some level of significance ( , P) These are a-priori decisions even when we don’t know what the data will be and how it will behave. Bayes: We get some evidence for the model (“likelihood”) and then can even compare “likelihoods” of different models

Where does Bayes Rule come at hand? In diagnostic cases where we’re are trying to calculate P(Disease | Symptom) we often know P(Symptom | Disease), the probability that you have the symptom given the disease, because this data has been collected from previous confirmed cases. In scientific cases where we want to know P(Hypothesis | Result), the probability that a hypothesis is true given some relevant result, we may know P(Result | Hypothesis), the probability that we would obtain that result given that the hypothesis is true- this is often statistically calculable, as when we have a p-value.

Applicability to (f)mri Let’s take fMRI as a relevant example Y = X *  +  We have: –Measured data : Y –Model : X –Model estimates: ,  (/variance)

What do we get with inferential statistics? T-statistics on the betas (  = (  1,  2,…)) (taking error into account) for a specific voxel  we would ONLY get that there is a chance (e.g.  2 ), given the data But what about the likelihood of the model??? What are the chances/likelihood that  1 >  2 at some voxel or region Could we get some quantitative measure on that?

What do we get with Bayes statistics? Here, the idea (Bayes) is to use our post-hoc knowledge (our data) to estimate the model, ( also allowing us to compare hypotheses (models) and see which fits our data best) “posterior” distribution for X given Y“likelihood” of Y given X prior probabilities of Y, X (“priors”) Now to Steve about the practical sides in SPM… P(X|Y) = P(Y|X)*P(X)/P(Y) i.e. P(  |Y) = P(Y|  )*P(  )/P(Y)

Bayes for Beginners: Applications

spatial normalization segmentation EEG source localisation and Bayesian inference in… Posterior Probability Maps (PPM) Dynamic Causal Modelling (DCM) SPM uses priors for estimation in… Bayes in SPM

 Standard approach in science is the null hypothesis significance test (NHST)  Low p value suggests “there is not nothing”  Assumption is H 0 = noise; randomness Null hypothesis significance testing H 0 = molecules are randomly arranged in space Looking unlikely… Kreuger (2001) American Psychologist

Something vs nothing …If there is any effect.. Our interpretations ultimately depend on p(H 0 ) “Risky” vs “safe” research… Better to be explicit – incorporate subjectivity when specifying hypotheses. Belief change = p(H 0 ) – p(H 0 | D) If the underlying effect δ ~= 0, no matter how small, the test statistic grows in size – is this physiological?

The case for the defence Law of large numbers means that the test statistic will identify a consistent trend (δ ~= 0) with a sufficient sample size In SPM, we look at images of statistics, not effect sizes A highly significant statistic may reflect a small non- physiological difference, with large N BUT… as long as we are aware of this, classical inference works well for common sample sizes

MpMp p -1 M post post -1 d -1 MdMd post = d + p M post = d M d + p M p post Posterior Probability Distribution precision = 1/  2

BUT!!!What is p(H 0 ) for randomness?! H1H1 H2H2 H3H3 H4H4 vs vsetc… Reframe the question – compare alternative hypotheses/models: (1) Bayesian model comparison Bayes: If only one model, then p(y) is a normalising constant… For model H i : Model evidence for H i

Practical example (1)

Dynamic causal modelling (DCM) V1 V5 SPC Motion Photic Attention H=1 V1 V5 SPC Motion Photic Attention H= V1 V5 SPC Motion Photic Attention H=3 Attention 0.03 Model Evidence: Bayes factor:

General Linear Model: What are the priors? with In “classical” SPM, no (flat) priors In “full” Bayes, priors might be from theoretical arguments or from independent data In “empirical” Bayes, priors derive from the same data, assuming a hierarchical model for generation of the data Parameters of one level can be made priors on distribution of parameters at lower level Parameters and hyperparameters at each level can be estimated using EM algorithm (2) Priors about the null hypothesis

Shrinkage prior General Linear Model: Shrinkage prior: 0 In the absence of evidence to the contrary, parameters will shrink to zero

Bayesian Inference Likelihood Prior Posterior SPMsSPMs PPMsPPMs Bayesian test Classical T-test Changes with search volume

Practical example (2)

SPM5 Interface

(2) Posterior Probability Maps Mean (Cbeta_*.img) Std dev (SDbeta_*.img) PPM (spmP_*.img) Activation threshold  Probability  Posterior probability distribution p(  |Y )

(3) Use informative priors (cutting edge!) Spatial constraints on fMRI activity (e.g. grey matter) Spatial constraints on EEG sources, e.g. using fMRI blobs ? ?

(4) Tasters – The Bayesian Brain

Ernst & Banks (2002) Nature (4a) Taster: Modelling behaviour…

Ernst & Banks (2002) Nature (4a) Taster: Modelling behaviour…

Friston (2005) Phil Trans R Soc B (4b) Taster: Modelling the brain…

Acknowledgements and further reading Previous MFD talks Jean & Guillame’s SPM course slides Krueger (2001) Null hypothesis significance testing Am Psychol 56: Penny et al. (2004) Comparing dynamic causal models. Neuroimage 22: Friston & Penny (2003) Posterior probability maps and SPMs Neuroimage 19: Friston (2005) A theory of cortical responses Phil Trans R Soc B

Bayes’ ending Bunhill Fields Burial Ground off City Road, EC1