University of Zurich, February 2011

Slides:



Advertisements
Similar presentations
Dynamic Causal Modelling (DCM) for fMRI
Advertisements

J. Daunizeau Institute of Empirical Research in Economics, Zurich, Switzerland Brain and Spine Institute, Paris, France Bayesian inference.
Bayesian inference Lee Harrison York Neuroimaging Centre 01 / 05 / 2009.
Bayesian models for fMRI data
DCM: Advanced Topics Klaas Enno Stephan Translational Neuromodeling Unit (TNU) Institute for Biomedical Engineering, University of Zurich & ETH Zurich.
DCM: Advanced topics Klaas Enno Stephan Laboratory for Social & Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
DCM: Advanced topics Klaas Enno Stephan SPM Course Zurich
Methods & Models for fMRI data analysis 17 December 2008
DCM: Advanced issues Klaas Enno Stephan Laboratory for Social & Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
DCM: Advanced issues Klaas Enno Stephan Laboratory for Social & Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
DCM: Advanced topics Klaas Enno Stephan Laboratory for Social & Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
DCM: Advanced topics Klaas Enno Stephan Laboratory for Social & Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
DCM: Advanced topics Klaas Enno Stephan Laboratory for Social & Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK Institute of Empirical Research in Economics, Zurich, Switzerland Bayesian inference.
Group analyses of fMRI data Methods & models for fMRI data analysis 26 November 2008 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
DCM: Advanced topics Rosalyn Moran Wellcome Trust Centre for Neuroimaging Institute of Neurology University College London With thanks to the FIL Methods.
GUIDE to The… D U M M I E S’ DCM Velia Cardin. Functional Specialization is a question of Where? Where in the brain is a certain cognitive/perceptual.
Dynamic Causal Modelling THEORY SPM Course FIL, London October 2009 Hanneke den Ouden Donders Centre for Cognitive Neuroimaging Radboud University.
Dynamic Causal Modelling
DCM Advanced, Part II Will Penny (Klaas Stephan) Wellcome Trust Centre for Neuroimaging Institute of Neurology University College London SPM Course 2014.
DCM: Advanced topics Klaas Enno Stephan Zurich SPM Course 2014
Dynamic Causal Modelling (DCM) for fMRI
Dynamic Causal Modelling (DCM): Theory Demis Hassabis & Hanneke den Ouden Thanks to Klaas Enno Stephan Functional Imaging Lab Wellcome Dept. of Imaging.
DCM for fMRI: Advanced topics Klaas Enno Stephan Laboratory for Social & Neural Systems Research (SNS) University of Zurich Wellcome Trust Centre for Neuroimaging.
Dynamic Causal Modelling (DCM) for fMRI
18 th February 2009 Stephanie Burnett Christian Lambert Methods for Dummies 2009 Dynamic Causal Modelling Part I: Theory.
Dynamic Causal Modelling Will Penny Wellcome Department of Imaging Neuroscience, University College London, UK FMRIB, Oxford, May
J. Daunizeau ICM, Paris, France ETH, Zurich, Switzerland Dynamic Causal Modelling of fMRI timeseries.
Dynamic Causal Modelling Advanced Topics SPM Course (fMRI), May 2015 Peter Zeidman Wellcome Trust Centre for Neuroimaging University College London.
Dynamic Causal Modelling (DCM) Marta I. Garrido Thanks to: Karl J. Friston, Klaas E. Stephan, Andre C. Marreiros, Stefan J. Kiebel,
Ch. 5 Bayesian Treatment of Neuroimaging Data Will Penny and Karl Friston Ch. 5 Bayesian Treatment of Neuroimaging Data Will Penny and Karl Friston 18.
Bayesian Model Comparison Will Penny London-Marseille Joint Meeting, Institut de Neurosciences Cognitive de la Mediterranee, Marseille, September 28-29,
Dynamic Causal Model for evoked responses in MEG/EEG Rosalyn Moran.
DCM: Advanced Topics Klaas Enno Stephan Translational Neuromodeling Unit (TNU) Institute for Biomedical Engineering University of Zurich & Swiss Federal.
DCM: Advanced issues Klaas Enno Stephan Centre for the Study of Social & Neural Systems Institute for Empirical Research in Economics University of Zurich.
Bayesian inference Lee Harrison York Neuroimaging Centre 23 / 10 / 2009.
Dynamic Causal Models Will Penny Olivier David, Karl Friston, Lee Harrison, Andrea Mechelli, Klaas Stephan Mathematics in Brain Imaging, IPAM, UCLA, USA,
Bayesian Inference in SPM2 Will Penny K. Friston, J. Ashburner, J.-B. Poline, R. Henson, S. Kiebel, D. Glaser Wellcome Department of Imaging Neuroscience,
Bayesian selection of dynamic causal models for fMRI Will Penny Olivier David, Karl Friston, Lee Harrison, Andrea Mechelli, Klaas Stephan The brain as.
Dynamic Causal Models Will Penny Olivier David, Karl Friston, Lee Harrison, Stefan Kiebel, Andrea Mechelli, Klaas Stephan MultiModal Brain Imaging, Copenhagen,
Bayesian Model Comparison Will Penny Wellcome Department of Imaging Neuroscience, University College London, UK UCL, June 7, 2003 Large-Scale Neural Network.
Bayesian Model Selection and Averaging SPM for MEG/EEG course Peter Zeidman 17 th May 2016, 16:15-17:00.
J. Daunizeau ICM, Paris, France TNU, Zurich, Switzerland
Group Analyses Guillaume Flandin SPM Course London, October 2016
DCM: Advanced Topics Klaas Enno Stephan SPM Course FIL London
5th March 2008 Andreina Mendez Stephanie Burnett
Dynamic Causal Modeling of Endogenous Fluctuations
Variational filtering in generated coordinates of motion
Effective Connectivity
Dynamic Causal Modelling (DCM): Theory
Computational models for imaging analyses
Wellcome Trust Centre for Neuroimaging University College London
Dynamic Causal Model for evoked responses in M/EEG Rosalyn Moran.
DCM: Advanced issues Klaas Enno Stephan Laboratory for Social & Neural Systems Research Institute for Empirical Research in Economics University of.
Dynamic Causal Modelling
SPM2: Modelling and Inference
Dynamic Causal Modelling for M/EEG
Dynamic Causal Modelling
Bayesian Methods in Brain Imaging
CRIS Workshop: Computational Neuroscience and Bayesian Modelling
Hierarchical Models and
Effective Connectivity
Bayesian inference J. Daunizeau
M/EEG Statistical Analysis & Source Localization
DCM for fMRI – Advanced Topics
Bayesian Inference in SPM2
Wellcome Centre for Neuroimaging, UCL, UK.
WellcomeTrust Centre for Neuroimaging University College London
Bayesian Model Selection and Averaging
Group DCM analysis for cognitive & clinical studies
Presentation transcript:

University of Zurich, 16-18 February 2011 DCM: Advanced topics Rosalyn Moran Wellcome Trust Centre for Neuroimaging Institute of Neurology University College London With thanks to the FIL Methods Group for slides and images SPM Course 2011 University of Zurich, 16-18 February 2011

Dynamic Causal Modeling (DCM) Hemodynamic forward model: neural activityBOLD Electromagnetic forward model: neural activityEEG MEG LFP Neural state equation: fMRI EEG/MEG simple neuronal model complicated forward model complicated neuronal model simple forward model inputs

Overview Bayesian model selection (BMS) Nonlinear DCM for fMRI Stochastic DCM Embedding computational models in DCMs Integrating tractography and DCM

Overview Bayesian model selection (BMS) Nonlinear DCM for fMRI Stochastic DCM Embedding computational models in DCMs Integrating tractography and DCM

Model comparison and selection Given competing hypotheses on structure & functional mechanisms of a system, which model is the best? Pitt & Miyung (2002) TICS Which model represents the best balance between model fit and model complexity? For which model m does p(y|m) become maximal?

Approximations to the model evidence in DCM Logarithm is a monotonic function Maximizing log model evidence = Maximizing model evidence Log model evidence = balance between fit and complexity No. of parameters In SPM2 & SPM5, interface offers 2 approximations: No. of data points Akaike Information Criterion: Bayesian Information Criterion: AIC favours more complex models, BIC favours simpler models. Penny et al. 2004, NeuroImage

The negative free energy approximation The negative free energy F is a lower bound on the log model evidence:

The complexity term in F In contrast to AIC & BIC, the complexity term of the negative free energy F accounts for parameter interdependencies. Under gaussian assumptions: The complexity term of F is higher the more independent the prior parameters ( effective DFs) the more dependent the posterior parameters the more the posterior mean deviates from the prior mean NB: SPM8 only uses F for model selection ! However, comparing these expression to Equation 4, shows that both the AIC and BIC will fail in various situations. An obvious example is redundant parameterisation; the true complexity will not change when we add a parameter whose effect is identical to another parameter in measurement space. While the free-energy bound would take this redundancy into account, keeping the complexity identical, the AIC and BIC approximations would indicate that complexity has increased. In practice, many models show partial dependencies amongst parameters, meaning that AIC and BIC routinely over-estimate the effect that adding or removing parameters has on model complexity Penny et al. submitted

Bayes factors For a given dataset, to compare two models, we compare their evidences. positive value, [0;[ B12 p(m1|y) Evidence 1 to 3 50-75% weak 3 to 20 75-95% positive 20 to 150 95-99% strong  150  99% Very strong Kass & Raftery classification: or their log evidences Kass & Raftery 1995, J. Am. Stat. Assoc.

BMS in SPM8: an example M1 M2 M3 M4 attention PPC PPC BF 2966 M2 better than M1 attention stim V1 V5 stim V1 V5 M1 M2 M3 M4 V1 V5 stim PPC M3 attention M3 better than M2 BF  12 F = 2.450 Posterior model probability in lower plot is a normalised probability: p(m_i|y) = p(y|m_i)/sum(p(y|m_i)) Note that under flat model priors p(m_i|y) = p(y|m_i) V1 V5 stim PPC M4 attention M4 better than M3 BF  23 F = 3.144

Fixed effects BMS at group level Group Bayes factor (GBF) for 1...K subjects: Average Bayes factor (ABF): Problems: blind with regard to group heterogeneity sensitive to outliers or

Random effects BMS for group studies Dirichlet parameters = “occurrences” of models in the population Dirichlet distribution of model probabilities Multinomial distribution of model labels Model inversion by Variational Bayes (VB) Want: the density from which models are sampled to generate subject-specific data. we seek the conditional estimates of the multinomial parameters, i.e. the model probabilities , that generate switches or indicator variables, , prescribing the model for the i-th subject; where . Since the model probabilities r follow a Dirichlet distribution , the conditional expectations encode the expected probability that the k-th model will be selected for any randomly selected subject. Also, the cumulative probability density of can be used to quantify our belief that a given model m1 has a higher probability of having generated the observed data across the group than a second model m2; below, we will call this the "exceedance probability". In the following, we describe a hierarchical Bayesian model that can be inverted to obtain the Dirichlet parameters . Using the Dirichlet density for model comparison After the above optimization of the Dirichlet parameters, , the Dirichlet density can be used for model comparisons at the group level. There are several ways to report the results of this comparison. The simplest option is to report the Dirichlet parameter estimates (c.f. Eq. 20). Another possibility is to use to compute the expected multinomial parameters and thus the expected likelihood of obtaining a particular model, i.e. for any randomly selected subject [1] (20) Either the Dirichlet parameter estimates  or the expected multinomial parameters can be used, with equivalent results, to rank models at the group level. A third option, which is perhaps the most useful one when two models (or model subsets, see below) are compared against each other, is to use the cumulative probability density of to quantify our belief that one model is more likely than a second one, given the group data. For example, when comparing two models, m1 and m2, the belief that m1 has the higher probability of having generated the observed data across the group corresponds to the "exceedance probability" [1] Note that for the special case of "drawing" a single "sample" (model), the multinomial distribution of models reduces to . Therefore, for any given subject, represents the conditional expectation that the k-th model generated her/his data. This enables us to estimate the Dirichlet parameters given observations or samples from a multinomial distribution We consider the same problem in Bayesian terms and describe a novel hierarchical model, which is optimised to furnish a probability density on the models themselves. This new method rests on a variational approach to estimating the parameters of a Dirichlet distribution, which define a multinomial distribution of how likely it is that a specific model generated the data from a specific subject. Using empirical and synthetic data, we show that optimising a conditional density of the model probabilities, given the log-evidences for each model over subjects the Dirichlet distribution is the conjugate prior of the categorical distribution and multinomial distribution. That is, its probability density function returns the belief that the probabilities of K rival events are xi given that each event has been observed αi − 1 times. The support of the Dirichlet distribution is a K-dimensional vector of real numbers in the range (0,1), all of which sum to 1. These can be viewed as the probabilities of a K-way categorical event. Another way to express this is that the domain of the Dirichlet distribution is itself a probability distribution, specifically a K-dimensional discrete distribution In probability theory, the multinomial distribution is a generalization of the binomial distribution. The binomial distribution is the probability distribution of the number of "successes" in n independent Bernoulli trials, with the same probability of "success" on each trial. In a multinomial distribution, the analog of the Bernoulli distribution is the categorical distribution, where each trial results in exactly one of some fixed finite number k of possible outcomes, with probabilities p1, ..., pk (so that pi ≥ 0 for i = 1, ..., k and ), and there are n independent trials. Then let the random variables Xi indicate the number of times outcome number i was observed over the n trials. The vector X = (X1, ..., Xk) follows a multinomial distribution with parameters n and p, where p = (p1, ..., pk). rm as the frequency with which model m is used in the population. Measured data y estimate the parameters  of the posterior Stephan et al. 2009, NeuroImage

Random effects BMS for group studies “the occurences” “the expected likelihood” “the exceedance probability” Alpha, the occurences, <r> the expected likelihood of obtaining a particular model for any randomly selected subject . For example, when comparing two models, m1 and m2, the belief that m1 has the higher probability of having generated the observed data across the group corresponds to the "exceedance probability Stephan et al. 2009, NeuroImage

group analysis (random effects), n=16, p<0.05 corrected Task-driven lateralisation Does the word contain the letter A or not? letter decisions > spatial decisions • group analysis (random effects), n=16, p<0.05 corrected analysis with SPM2 time Is the red letter left or right from the midline of the word? spatial decisions > letter decisions Stephan et al. 2003, Science

+ Theories on inter-hemispheric integration during lateralised tasks Information transfer (for left-lateralised task) T |RVF  + T |LVF LVF RVF Predictions: modulation by task conditional on visual field asymmetric connection strengths

Ventral stream & letter decisions Left MOG -38,-90,-4 Left FG -44,-52,-18 Right FG 38,-52,-20 Right MOG -38,-94,0 LD|LVF LD>SD, p<0.05 cluster-level corrected (p<0.001 voxel-level cut-off) p<0.01 uncorrected MOG left FG left FG right MOG right LG left LG right Left LG -12,-70,-6 Left LG -14,-68,-2 RVF stim. LVF stim. LD>SD masked incl. with RVF>LVF p<0.05 cluster-level corrected (p<0.001 voxel-level cut-off) LD>SD masked incl. with LVF>RVF p<0.05 cluster-level corrected (p<0.001 voxel-level cut-off) Stephan et al. 2007, J. Neurosci.

Ventral stream & letter decisions Left MOG -38,-90,-4 Left FG -44,-52,-18 Right FG 38,-52,-20 Right MOG -38,-94,0 LD>SD, p<0.05 cluster-level corrected (p<0.001 voxel-level cut-off) p<0.01 uncorrected MOG left FG left FG right MOG right LG left LG right Left LG -12,-70,-6 Left LG -14,-68,-2 LD|LVF RVF stim. LVF stim. LD>SD masked incl. with RVF>LVF p<0.05 cluster-level corrected (p<0.001 voxel-level cut-off) LD>SD masked incl. with LVF>RVF p<0.05 cluster-level corrected (p<0.001 voxel-level cut-off) Stephan et al. 2007, J. Neurosci.

m2 m1 m2 m1 Winner! Fixed Effects Stephan et al. 2009, NeuroImage MOG LG RVF stim. LVF FG LD LD|RVF LD|LVF MOG LG RVF stim. LVF FG LD|RVF LD|LVF LD m1 m2 m1 Stephan et al. 2009, NeuroImage

m2 m1

Simulation study: sampling subjects from a heterogenous population MOG LG RVF stim. LVF FG LD|RVF LD|LVF LD m1 Population where 70% of all subjects' data are generated by model m1 and 30% by model m2 Random sampling of subjects from this population and generating synthetic data with observation noise Fitting both m1 and m2 to all data sets and performing BMS MOG LG RVF stim. LVF FG LD LD|RVF LD|LVF m2 Stephan et al. 2009, NeuroImage

 <r>  m1 m2 m1 m2 m1 m2 true values: 1=220.7=15.4 2=220.3=6.6 mean estimates: 1=15.4, 2=6.6 true values: r1 = 0.7, r2=0.3 mean estimates:  <r> m1 m2 m1 m2 true values: 1 = 1, 2=0 mean estimates: 1 = 0.89, 2=0.11  m1 m2

Families of Models Partition the cortical dynamics of intelligible Speech activity among three key multimodal regions: the left posterior and anterior superior temporal sulcus (subsequently referred to as regions P and A respectively) and pars orbitalis of the inferior frontal gyrus (region F). The aim of the study was to see how connections among regions depended on whether the auditory input was intelligible speech or time-reversed speech. Partition

Families of Models * * e.g. Modulatory connections BMA: weight posterior parameter densities with model probabilities Penny et al., 2010

definition of model space inference on model structure or inference on model parameters? inference on individual models or model space partition? inference on parameters of an optimal model or parameters of all models? optimal model structure assumed to be identical across subjects? comparison of model families using FFX or RFX BMS optimal model structure assumed to be identical across subjects? BMA yes no yes no FFX BMS RFX BMS FFX BMS RFX BMS FFX analysis of parameter estimates (e.g. BPA) RFX analysis of parameter estimates (e.g. t-test, ANOVA) Stephan et al. 2010, NeuroImage

Overview Bayesian model selection (BMS) Nonlinear DCM for fMRI Stochastic DCM Embedding computational models in DCMs Integrating tractography and DCM

   BOLD y y y y λ x neuronal states hemodynamic model activity x2(t) activity x3(t) activity x1(t) x neuronal states modulatory input u2(t) t integration intrinsic connectivity direct inputs modulation of connectivity Neural state equation t driving input u1(t) Stephan & Friston (2007), Handbook of Brain Connectivity

bilinear DCM non-linear DCM driving input modulation driving input modulation Two-dimensional Taylor series (around x0=0, u0=0): Bilinear state equation: Nonlinear state equation:

u2 x1 x2 x3 u1 Nonlinear dynamic causal model (DCM): Neural population activity fMRI signal change (%) u2 x1 x2 x3 u1 Nonlinear dynamic causal model (DCM): Stephan et al. 2008, NeuroImage

Nonlinear DCM: Attention to motion Stimuli + Task Previous bilinear DCM V1 IFG V5 SPC Motion Photic Attention .82 (100%) .42 .37 (90%) .69 (100%) .47 .65 (100%) .52 (98%) .56 (99%) Büchel & Friston (1997) 250 radially moving dots (4.7 °/s) Friston et al. (2003) Conditions: F – fixation only A – motion + attention (“detect changes”) N – motion without attention S – stationary dots Friston et al. (2003): attention modulates backward connections IFG→SPC and SPC→V5. Q: Is a nonlinear mechanism (gain control) a better explanation of the data?

M1 M2   M3  M4 modulation of back- ward or forward connection? attention M1 M2  modulation of back- ward or forward connection? PPC PPC BF = 2966 M2 better than M1 attention stim V1 V5 stim V1 V5 M3 better than M2 BF = 12  additional driving effect of attention on PPC? V1 V5 stim PPC M3 attention M4 better than M3 BF = 23  bilinear or nonlinear modulation of forward connection? V1 V5 stim PPC M4 attention Stephan et al. 2008, NeuroImage

attention PPC stim V1 V5 motion MAP = 1.25 0.10 PPC 0.26 0.39 1.25 0.26 stim V1 0.13 V5 0.46 0.50 motion Stephan et al. 2008, NeuroImage

motion & attention motion & no attention static dots V1 V5 PPC observed fitted

Overview Bayesian model selection (BMS) Nonlinear DCM for fMRI Stochastic DCM Embedding computational models in DCMs Integrating tractography and DCM % Generalised filtering (under the Laplace assumption) % ===================================================================== if DCM.options.stochastic == 1 DEM = spm_LAP(DEM); % no mean-field assumption Stochastic DCMs: Accounting for stochastic inputs to the network and their interaction with task-specific processes may be of particular importance for studying state-dependent processes, e.g., short-term plasticity and trial-by-trial variations of effective connectivity. In addition, provided that the probabilistic inversion schemes are properly extended

Stochastic DCMs   neuronal states Daunizeau et al, 2009 Stochastic innovations: variance hyperparameter   activity x2(t) activity x3(t) activity x1(t) neuronal states modulatory input u2(t) t t driving input u1(t) Daunizeau et al, 2009 Friston et al, 2008 Inversion: Generalised filtering (under the Laplace assumption)

Overview Bayesian model selection (BMS) Nonlinear DCM for fMRI Stochastic DCM Embedding computational models in DCMs Integrating tractography and DCM

Learning of dynamic audio-visual associations 200 400 600 800 1000 0.2 0.4 0.6 0.8 1 CS 1 2 CS Response Time (ms) 200 400 600 800 2000 ± 650 or Target Stimulus Conditioning Stimulus TS p(face) Specifically, the cue could be (1) strongly predictive ( p0.9), (2) predictive ( p 0.7), (3) nonpredictive ( p 0.5), (4) antipredictive ( p 0.3), and (5) strongly antipredictive ( p 0.1) of the visual stimulus. trial den Ouden et al. 2010, J. Neurosci .

Bayesian learning model k vt-1 vt rt rt+1 ut ut+1 volatility probabilistic association observed events First, we assessed the main effect of probability, that is, in which in brain regions the hemodynamic response reflected the probability of the stimulus occurring, independently of which stimulus it was. We tested for both BOLD responses that increased with the likelihood of the outcome and responses that increased the less likely (or more surprising) the outcome was. In other words, these contrasts tested for stimulusindependent responses that reflected predicted or surprising outcomes, respectively. Given the results from our previous study (den Ouden et al., 2009), our a priori hypothesis was that the response in the putamen would correlate positively with prediction error, i.e., negatively with the probability of the observed outcome. Changes over trials: Model Based Regressor Behrens et al. 2007, Nat. Neurosci.

Comparison with competing learning models 400 440 480 520 560 600 0.2 0.4 0.6 0.8 1 Trial p(F) True Bayes Vol HMM fixed HMM learn RW Alternative learning models: Rescorla-Wagner HMM (2 variants) True probabilities BMS: hierarchical Bayesian learner performs best The posterior mean of p(FCS) as estimated by the Bayesian learner (dashed line) tracks the underlying blocked probabilities (solid line). Note that the blocked probabilities are CS2 zoomed in from B (trials 400–600, session 3). Because blocks of stable probabilities are short, however, the estimated probabilities never quite reach their true values during a given block. Note that the estimates change rapidly at block transitions. When an unexpected stimulus occurs, the estimates briefly move toward p0.5. Note den Ouden et al. 2010, J. Neurosci .

Stimulus-independent prediction error Putamen Premotor cortex p < 0.05 (SVC) p < 0.05 (cluster-level whole- brain corrected) p(F) p(H) -2 -1.5 -1 -0.5 BOLD resp. (a.u.) p(F) p(H) -2 -1.5 -1 -0.5 BOLD resp. (a.u.) den Ouden et al. 2010, J. Neurosci .

Prediction error (PE) activity in the putamen PE during reinforcement learning O'Doherty et al. 2004, Science PE during incidental sensory learning den Ouden et al. 2009, Cerebral Cortex According to the free energy principle (and other learning theories): synaptic plasticity during learning = PE dependent changes in connectivity

Prediction error in PMd: cause or effect? Model 1 Model 2 den Ouden et al. 2010, J. Neurosci .

Prediction error gates visuo-motor connections Modulation of visuo-motor connections by striatal PE activity Influence of visual areas on premotor cortex: stronger for surprising stimuli weaker for expected stimuli p(H) p(F) PUT d = 0.010 0.003 p = 0.010 d = 0.011 0.004 p = 0.017 PMd PPA FFA den Ouden et al. 2010, J. Neurosci .

Overview Bayesian model selection (BMS) Nonlinear DCM for fMRI Stochastic DCM Embedding computational models in DCMs Integrating tractography and DCM

Diffusion-tensor imaging Parker & Alexander, 2005, Phil. Trans. B

Probabilistic tractography: Kaden et al. 2007, NeuroImage computes local fibre orientation density by deconvolution of the diffusion-weighted signal estimates the spatial probability distribution of connectivity from given seed regions anatomical connectivity = proportion of fibre pathways originating in a specific source region that intersect a target region If the area or volume of the source region approaches a point, this measure reduces to method by Behrens et al. (2003)

Integration of tractography and DCM low probability of anatomical connection  small prior variance of effective connectivity parameter R1 R2 high probability of anatomical connection  large prior variance of effective connectivity parameter Stephan, Tittgemeyer et al. 2009, NeuroImage

 anatomical connectivity  DCM structure LD|LVF probabilistic tractography LG left right FG  anatomical connectivity FG (x3) FG (x4) LD LD LG (x1) LG (x2)  DCM structure LD|RVF RVF stim. BVF stim. LVF stim.  connection-specific priors for coupling parameters Stephan, Tittgemeyer et al. 2009, NeuroImage

Connection-specific prior variance  as a function of anatomical connection probability  64 different mappings by systematic search across hyper-parameters  and  yields anatomically informed (intuitive and counterintuitive) and uninformed priors

Stephan, Tittgemeyer et al. 2009, NeuroImage

Methods papers on DCM for fMRI and BMS – part 1 Daunizeau J., Friston K. J., Kiebel S. J. ; Variational Bayesian identification and prediction of stochastic nonlinear dynamic causal models, Physica D (2009) 238: 2089-2118. Chumbley JR, Friston KJ, Fearn T, Kiebel SJ (2007) A Metropolis-Hastings algorithm for dynamic causal models. Neuroimage 38:478-487. Daunizeau J, David, O, Stephan KE (2010) Dynamic Causal Modelling: A critical review of the biophysical and statistical foundations. NeuroImage, in press. Friston KJ, Harrison L, Penny W (2003) Dynamic causal modelling. NeuroImage 19:1273-1302. Kasess CH, Stephan KE, Weissenbacher A, Pezawas L, Moser E, Windischberger C (2010) Multi-Subject Analyses with Dynamic Causal Modeling. NeuroImage 49: 3065-3074. Kiebel SJ, Kloppel S, Weiskopf N, Friston KJ (2007) Dynamic causal modeling: a generative model of slice timing in fMRI. NeuroImage 34:1487-1496. Marreiros AC, Kiebel SJ, Friston KJ (2008) Dynamic causal modelling for fMRI: a two-state model. NeuroImage 39:269-278. Penny WD, Stephan KE, Mechelli A, Friston KJ (2004a) Comparing dynamic causal models. NeuroImage 22:1157-1172. Penny WD, Stephan KE, Mechelli A, Friston KJ (2004b) Modelling functional integration: a comparison of structural equation and dynamic causal models. NeuroImage 23 Suppl 1:S264- 274. Penny WD, Stephan KE, Daunizeau J, Joao M, Friston K, Schofield T, Leff AP (2010) Comparing Families of Dynamic Causal Models. PLoS Computational Biology, in press.

Methods papers on DCM for fMRI and BMS – part 2 Stephan KE, Harrison LM, Penny WD, Friston KJ (2004) Biophysical models of fMRI responses. Curr Opin Neurobiol 14:629-635. Stephan KE, Weiskopf N, Drysdale PM, Robinson PA, Friston KJ (2007) Comparing hemodynamic models with DCM. NeuroImage 38:387-401. Stephan KE, Harrison LM, Kiebel SJ, David O, Penny WD, Friston KJ (2007) Dynamic causal models of neural system dynamics: current state and future extensions. J Biosci 32:129-144. Stephan KE, Weiskopf N, Drysdale PM, Robinson PA, Friston KJ (2007) Comparing hemodynamic models with DCM. Neuroimage 38:387-401. Stephan KE, Kasper L, Harrison LM, Daunizeau J, den Ouden HE, Breakspear M, Friston KJ (2008) Nonlinear dynamic causal models for fMRI. NeuroImage 42:649-662. Stephan KE, Penny WD, Daunizeau J, Moran RJ, Friston KJ (2009) Bayesian model selection for group studies. NeuroImage 46:1004-1017. Stephan KE, Tittgemeyer M, Knösche TR, Moran RJ, Friston KJ (2009) Tractography-based priors for dynamic causal models. NeuroImage 47: 1628-1638. Stephan KE, Penny WD, Moran RJ, den Ouden HEM, Daunizeau J, Friston KJ (2010) Ten simple rules for Dynamic Causal Modelling. NeuroImage 49: 3099-3109.

Thank you