Karl Friston, Oliver Josephs

Slides:



Advertisements
Similar presentations
Basis Functions. What’s a basis ? Can be used to describe any point in space. e.g. the common Euclidian basis (x, y, z) forms a basis according to which.
Advertisements

The General Linear Model Or, What the Hell’s Going on During Estimation?
1 Temporal Processing Chris Rorden Temporal Processing can reduce error in our model –Slice Time Correction –Temporal Autocorrelation –High and low pass.
1 Temporal Processing Chris Rorden Temporal Processing can reduce error in our model –Slice Time Correction –Temporal Autocorrelation –High and low pass.
Event-related fMRI Will Penny (this talk was made by Rik Henson) Event-related fMRI Will Penny (this talk was made by Rik Henson)
Efficiency in Experimental Design Catherine Jones MfD2004.
Event-related fMRI (er-fMRI) Methods & models for fMRI data analysis 25 March 2009 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
The General Linear Model (GLM) Methods & models for fMRI data analysis in neuroeconomics November 2010 Klaas Enno Stephan Laboratory for Social & Neural.
Event-related fMRI (er-fMRI) Methods & models for fMRI data analysis 05 November 2008 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
The General Linear Model (GLM)
The General Linear Model (GLM) SPM Course 2010 University of Zurich, February 2010 Klaas Enno Stephan Laboratory for Social & Neural Systems Research.
1st level analysis: basis functions and correlated regressors
Parametric modulation, temporal basis functions and correlated regressors Mkael Symmonds Antoinette Nicolle Methods for Dummies 21 st January 2008.
SPM short course – May 2003 Linear Models and Contrasts The random field theory Hammering a Linear Model Use for Normalisation T and F tests : (orthogonal.
With many thanks for slides & images to: FIL Methods group, Virginia Flanagin and Klaas Enno Stephan Dr. Frederike Petzschner Translational Neuromodeling.
Contrasts (a revision of t and F contrasts by a very dummyish Martha) & Basis Functions (by a much less dummyish Iroise!)
SPM short course – Oct Linear Models and Contrasts Jean-Baptiste Poline Neurospin, I2BM, CEA Saclay, France.
Event-related fMRI Rik Henson With thanks to: Karl Friston, Oliver Josephs Event-related fMRI Rik Henson With thanks to: Karl Friston, Oliver Josephs.
Wellcome Dept. of Imaging Neuroscience University College London
(Epoch &) Event-related fMRI SPM Course 2002 (Epoch &) Event-related fMRI SPM Course 2002 Christian Buechel Karl Friston Rik Henson Oliver Josephs Wellcome.
(Epoch and) Event-related fMRI Karl Friston Rik Henson Oliver Josephs Wellcome Department of Cognitive Neurology & Institute of Cognitive Neuroscience.
Event-related fMRI Rik Henson With thanks to: Karl Friston, Oliver Josephs Event-related fMRI Rik Henson With thanks to: Karl Friston, Oliver Josephs.
Spatial smoothing of autocorrelations to control the degrees of freedom in fMRI analysis Keith Worsley Department of Mathematics and Statistics, McGill.
The General Linear Model (for dummies…) Carmen Tur and Ashwani Jha 2009.
Temporal Basis Functions Melanie Boly Methods for Dummies 27 Jan 2010.
1 Time Series Analysis of fMRI II: Noise, Inference, and Model Error Douglas N. Greve
Event-related fMRI SPM course May 2015 Helen Barron Wellcome Trust Centre for Neuroimaging 12 Queen Square.
FMRI Modelling & Statistical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course Chicago, Oct.
The General Linear Model
The General Linear Model Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM fMRI Course London, May 2012.
SPM short – Mai 2008 Linear Models and Contrasts Stefan Kiebel Wellcome Trust Centre for Neuroimaging.
The general linear model and Statistical Parametric Mapping I: Introduction to the GLM Alexa Morcom and Stefan Kiebel, Rik Henson, Andrew Holmes & J-B.
Event-related fMRI Christian Ruff With thanks to: Rik Henson.
The general linear model and Statistical Parametric Mapping II: GLM for fMRI Alexa Morcom and Stefan Kiebel, Rik Henson, Andrew Holmes & J-B Poline.
The General Linear Model Christophe Phillips SPM Short Course London, May 2013.
The General Linear Model Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM fMRI Course London, October 2012.
Covariance components II autocorrelation & nonsphericity
The General Linear Model (GLM)
Group Analyses Guillaume Flandin SPM Course London, October 2016
The General Linear Model (GLM)
Statistical Parametric
Variational Bayesian Inference for fMRI time series
The general linear model and Statistical Parametric Mapping
The General Linear Model
Karl Friston, Oliver Josephs
Karl Friston, Oliver Josephs
The General Linear Model (GLM): the marriage between linear systems and stats FFA.
SPM short course at Yale – April 2005 Linear Models and Contrasts
and Stefan Kiebel, Rik Henson, Andrew Holmes & J-B Poline
Barbora Ondrejickova Methods for Dummies 23rd November 2016
The SPM MfD course 12th Dec 2007 Elvina Chu
The General Linear Model (GLM)
Sam Ereira Methods for Dummies 13th January 2016
The General Linear Model
Karl Friston, Oliver Josephs
The general linear model and Statistical Parametric Mapping
SPM2: Modelling and Inference
Bayesian Methods in Brain Imaging
The General Linear Model
Hierarchical Models and
The General Linear Model (GLM)
Event-related fMRI Christian Ruff With thanks to: Rik Henson.
Chapter 3 General Linear Model
MfD 04/12/18 Alice Accorroni – Elena Amoruso
Bayesian Inference in SPM2
The General Linear Model
The General Linear Model (GLM)
Probabilistic Modelling of Brain Imaging Data
The General Linear Model
The General Linear Model
Presentation transcript:

Karl Friston, Oliver Josephs Convolution Models for fMRI Rik Henson With thanks to: Karl Friston, Oliver Josephs

Overview 1. BOLD impulse response 2. Convolution and GLM 3. Temporal Basis Functions 4. Filtering/Autocorrelation 5. Timing Issues 6. Nonlinear Models

BOLD Impulse Response Function of blood oxygenation, flow, volume (Buxton et al, 1998) Peak (max. oxygenation) 4-6s poststimulus; baseline after 20-30s Initial undershoot can be observed (Malonek & Grinvald, 1996) Similar across V1, A1, S1… … but differences across: other regions (Schacter et al 1997) individuals (Aguirre et al, 1998) Brief Stimulus Undershoot Initial Peak

BOLD Impulse Response Early event-related fMRI studies used a long Stimulus Onset Asynchrony (SOA) to allow BOLD response to return to baseline However, if the BOLD response is explicitly modelled, overlap between successive responses at short SOAs can be accommodated… … particularly if responses are assumed to superpose linearly Short SOAs are more sensitive… Brief Stimulus Undershoot Initial Peak

Overview 1. BOLD impulse response 2. Convolution and GLM 3. Temporal Basis Functions 4. Filtering/Autocorrelation 5. Timing Issues 6. Nonlinear Models

General Linear (Convolution) Model GLM for a single voxel: y(t) = u(t)  h(t) + (t) u(t) = neural causes (stimulus train) u(t) =   (t - nT) h(t) = hemodynamic (BOLD) response h(t) =  ßi fi (t) fi(t) = temporal basis functions y(t) =   ßi fi (t - nT) + (t) y = X ß + ε T 2T 3T ... u(t) h(t)= ßi fi (t) convolution sampled each scan Design Matrix

General Linear Model (in SPM) Auditory words every 20s SPM{F} 0 time {secs} 30 Gamma functions ƒi() of peristimulus time  (Orthogonalised) Sampled every TR = 1.7s Design matrix, X [x(t)ƒ1() | x(t)ƒ2() |...] …

A word about down-sampling T=16, TR=2s Scan 1 o T0=16 o T0=9 x2 x3

Overview 1. BOLD impulse response 2. Convolution and GLM 3. Temporal Basis Functions 4. Filtering/Autocorrelation 5. Timing Issues 6. Nonlinear Models

Temporal Basis Functions Fourier Set Windowed sines & cosines Any shape (up to frequency limit) Inference via F-test

Temporal Basis Functions Finite Impulse Response Mini “timebins” (selective averaging) Any shape (up to bin-width) Inference via F-test

Temporal Basis Functions Fourier Set Windowed sines & cosines Any shape (up to frequency limit) Inference via F-test Gamma Functions Bounded, asymmetrical (like BOLD) Set of different lags

Temporal Basis Functions Fourier Set Windowed sines & cosines Any shape (up to frequency limit) Inference via F-test Gamma Functions Bounded, asymmetrical (like BOLD) Set of different lags “Informed” Basis Set Best guess of canonical BOLD response Variability captured by Taylor expansion “Magnitude” inferences via t-test…?

Temporal Basis Functions

Temporal Basis Functions “Informed” Basis Set (Friston et al. 1998) Canonical HRF (2 gamma functions) Canonical

Temporal Basis Functions “Informed” Basis Set (Friston et al. 1998) Canonical HRF (2 gamma functions) plus Multivariate Taylor expansion in: time (Temporal Derivative) Canonical Temporal

Temporal Basis Functions “Informed” Basis Set (Friston et al. 1998) Canonical HRF (2 gamma functions) plus Multivariate Taylor expansion in: time (Temporal Derivative) width (Dispersion Derivative) Canonical Temporal Dispersion

Temporal Basis Functions “Informed” Basis Set (Friston et al. 1998) Canonical HRF (2 gamma functions) plus Multivariate Taylor expansion in: time (Temporal Derivative) width (Dispersion Derivative) F-tests allow for “canonical-like” responses Canonical Temporal Dispersion

Temporal Basis Functions “Informed” Basis Set (Friston et al. 1998) Canonical HRF (2 gamma functions) plus Multivariate Taylor expansion in: time (Temporal Derivative) width (Dispersion Derivative) F-tests allow for any “canonical-like” responses T-tests on canonical HRF alone (at 1st level) can be improved by derivatives reducing residual error, and can be interpreted as “amplitude” differences, assuming canonical HRF is good fit… Canonical Temporal Dispersion

(Other Approaches) Long Stimulus Onset Asychrony (SOA) Can ignore overlap between responses (Cohen et al 1997) … but long SOAs are less sensitive Fully counterbalanced designs Assume response overlap cancels (Saykin et al 1999) Include fixation trials to “selectively average” response even at short SOA (Dale & Buckner, 1997) … but unbalanced when events defined by subject Define HRF from pilot scan on each subject May capture intersubject variability (Zarahn et al, 1997) … but not interregional variability Numerical fitting of highly parametrised response functions Separate estimate of magnitude, latency, duration (Kruggel et al 1999) … but computationally expensive for every voxel

Temporal Basis Sets: Which One? In this example (rapid motor response to faces, Henson et al, 2001)… Canonical + Temporal + Dispersion + FIR …canonical + temporal + dispersion derivatives appear sufficient …may not be for more complex trials (eg stimulus-delay-response) …but then such trials better modelled with separate neural components (ie activity no longer delta function) + constrained HRF (Zarahn, 1999)

Overview 1. BOLD impulse response 2. Convolution and GLM 3. Temporal Basis Functions 4. Filtering/Autocorrelation 5. Timing Issues 6. Nonlinear Models

Low frequency noise Low frequency noise: Physical (scanner drifts) aliasing Low frequency noise: Physical (scanner drifts) Physiological (aliased) cardiac (~1 Hz) respiratory (~0.25 Hz) power spectrum highpass filter power spectrum noise signal (eg infinite 30s on-off)

Temporal autocorrelation… Because the data are typically correlated from one scan to the next, one cannot assume the degrees of freedom (dfs) are simply the number of scans minus the dfs used in the model – need “effective degrees of freedom” In other words, the residual errors are not independent: Y = Xb + e e ~ N(0,s2V) V  I, V=AA' where A is the intrinsic autocorrelation Generalised least squares: KY = KXb + Ke Ke ~ N(0, s2V) V = KAA'K' (autocorrelation is a special case of “nonsphericity”…)

Temporal autocorrelation (History) KY = KXb + Ke Ke ~ N(0, s2V) V = KAA'K' One method is to estimate A, using, for example, an AR(p) model, then: K = A-1 V = I (allows OLS) This “pre-whitening” is sensitive, but can be biased if K mis-estimated Another method (SPM99) is to smooth the data with a known autocorrelation that swamps any intrinsic autocorrelation: K = S V = SAA'S’ ~ SS' (use GLS) Effective degrees of freedom calculated with Satterthwaite approximation ( df = trace(RV)2/trace(RVRV) ) This is more robust (providing the temporal smoothing is sufficient, eg 4s FWHM Gaussian), but less sensitive Most recent method (SPM2/5) is to restrict K to highpass filter, and estimate residual autocorrelation A using voxel-wide, one-step ReML…

Nonsphericity and ReML (SPM2/SPM5) Scans cov(e) spherical Nonsphericity means (kind of) that: Ce = cov(e)  s2I Nonsphericity can be modelled by set of variance components: Ce = 1Q1 + 2Q2 + 3Q3 ... (i are hyper-parameters) - Non-identical (inhomogeneous): (e.g, two groups of subjects) - Non-independent (autocorrelated): (e.g, white noise + AR(1))

Nonsphericity and ReML (SPM2/SPM5) Joint estimation of parameters and hyper-parameters requires ReML ReML gives (Restricted) Maximum Likelihood (ML) estimates of (hyper)parameters, rather than Ordinary Least Square (OLS) estimates ML estimates are more efficient, entail exact dfs (no Satterthwaite approx)… …but computationally expensive: ReML is iterative (unless only one hyper-parameter) To speed up: Correlation of errors (V) estimated by pooling over voxels Covariance of errors (s2V) estimated by single, voxel-specific scaling hyperparameter Ce = ReML( yyT, X, Q ) ^ bOLS = (XTX)-1XTy (= X+y) bML = (XTCe-1X)-1XTCe-1y V = ReML(  yjyjT, X, Q )

Overview 1. BOLD impulse response 2. Convolution and GLM 3. Temporal Basis Functions 4. Filtering/Autocorrelation 5. Timing Issues 6. Nonlinear Models

Timing Issues : Practical TR=4s Scans Typical TR for 48 slice EPI at 3mm spacing is ~ 4s

Timing Issues : Practical TR=4s Scans Typical TR for 48 slice EPI at 3mm spacing is ~ 4s Sampling at [0,4,8,12…] post- stimulus may miss peak signal Stimulus (synchronous) SOA=8s Sampling rate=4s

Timing Issues : Practical TR=4s Scans Typical TR for 48 slice EPI at 3mm spacing is ~ 4s Sampling at [0,4,8,12…] post- stimulus may miss peak signal Higher effective sampling by: 1. Asynchrony eg SOA=1.5TR Stimulus (asynchronous) SOA=6s Sampling rate=2s

Timing Issues : Practical TR=4s Scans Typical TR for 48 slice EPI at 3mm spacing is ~ 4s Sampling at [0,4,8,12…] post- stimulus may miss peak signal Higher effective sampling by: 1. Asynchrony eg SOA=1.5TR 2. Random Jitter eg SOA=(2±0.5)TR Stimulus (random jitter) Sampling rate=2s

Timing Issues : Practical TR=4s Scans Typical TR for 48 slice EPI at 3mm spacing is ~ 4s Sampling at [0,4,8,12…] post- stimulus may miss peak signal Higher effective sampling by: 1. Asynchrony eg SOA=1.5TR 2. Random Jitter eg SOA=(2±0.5)TR Better response characterisation (Miezin et al, 2000) Stimulus (random jitter) Sampling rate=2s

Timing Issues : Practical …but “Slice-timing Problem” (Henson et al, 1999) Slices acquired at different times, yet model is the same for all slices

Timing Issues : Practical Top Slice Bottom Slice …but “Slice-timing Problem” (Henson et al, 1999) Slices acquired at different times, yet model is the same for all slices => different results (using canonical HRF) for different reference slices TR=3s SPM{t} SPM{t}

Timing Issues : Practical Top Slice Bottom Slice …but “Slice-timing Problem” (Henson et al, 1999) Slices acquired at different times, yet model is the same for all slices => different results (using canonical HRF) for different reference slices Solutions: 1. Temporal interpolation of data … but less good for longer TRs TR=3s SPM{t} SPM{t} Interpolated SPM{t}

Timing Issues : Practical Top Slice Bottom Slice …but “Slice-timing Problem” (Henson et al, 1999) Slices acquired at different times, yet model is the same for all slices => different results (using canonical HRF) for different reference slices Solutions: 1. Temporal interpolation of data … but less good for longer TRs 2. More general basis set (e.g., with temporal derivatives) … use a composite estimate, or use F-tests TR=3s SPM{t} SPM{t} Interpolated SPM{t} Derivative SPM{F}

Timing Issues : Latency Assume the real response, r(t), is a scaled (by ) version of the canonical, f(t), but delayed by a small amount dt: r(t) =  f(t+dt) ~  f(t) +  f ´(t) dt 1st-order Taylor If the fitted response, R(t), is modelled by the canonical + temporal derivative: R(t) = ß1 f(t) + ß2 f ´(t) GLM fit If want to reduce estimate of BOLD impulse response to one composite value, with some robustness to latency issues (e.g, real, or induced by slice-timing):  = sqrt(ß12 + ß22) (Calhoun et al, 2004) (similar logic applicable to other partial derivatives)

Timing Issues : Latency Assume the real response, r(t), is a scaled (by ) version of the canonical, f(t), but delayed by a small amount dt: r(t) =  f(t+dt) ~  f(t) +  f ´(t) dt 1st-order Taylor If the fitted response, R(t), is modelled by the canonical + temporal derivative: R(t) = ß1 f(t) + ß2 f ´(t) GLM fit …or if want to estimate latency directly (assuming 1st-order approx holds):  ~ ß1 dt ~ ß2 / ß1 (Henson et al, 2002) (Liao et al, 2002) ie, Latency can be approximated by the ratio of derivative-to-canonical parameter estimates (within limits of first-order approximation, +/-1s)

Overview 1. BOLD impulse response 2. Convolution and GLM 3. Temporal Basis Functions 4. Filtering/Autocorrelation 5. Timing Issues 6. Nonlinear Models

Nonlinear Model [u(t)] Volterra series - a general nonlinear input-output model y(t) = 1[u(t)] + 2[u(t)] + .... + n[u(t)] + .... n[u(t)] = ....  hn(t1,..., tn)u(t - t1) .... u(t - tn)dt1 .... dtn [u(t)] response y(t) input u(t) Stimulus function kernels (h) estimate

Nonlinear Model Friston et al (1997) kernel coefficients - h SPM{F} SPM{F} testing H0: kernel coefficients, h = 0

Nonlinear Model Friston et al (1997) kernel coefficients - h SPM{F} SPM{F} testing H0: kernel coefficients, h = 0 Significant nonlinearities at SOAs 0-10s: (e.g., underadditivity from 0-5s)

Underadditivity at short SOAs Nonlinear Effects Underadditivity at short SOAs Linear Prediction Volterra Prediction

Underadditivity at short SOAs Nonlinear Effects Underadditivity at short SOAs Linear Prediction Volterra Prediction

Underadditivity at short SOAs Nonlinear Effects Underadditivity at short SOAs Linear Prediction Implications for Efficiency Volterra Prediction

The End This talk appears as Chapter 14 in the SPM book: http://www.mrc-cbu.cam.ac.uk/~rh01/Henson_Conv_SPMBook_2006_preprint.pdf

Temporal Basis Sets: Which One? Group Analysis (N=12) Null space of Canonical & Derivatives Null space of Canonical

Timing Issues : Latency Shifted Responses (green/ yellow) Canonical Canonical Derivative Basis Functions ß1 ß2 Parameter Estimates ß2 /ß1 Actual latency, dt, vs. ß2 / ß1 Face repetition reduces latency as well as magnitude of fusiform response

Timing Issues : Latency Neural BOLD A. Decreased B. Advanced C. Shortened (same integrated) D. Shortened (same maximum) A. Smaller Peak B. Earlier Onset C. Earlier Peak D. Smaller Peak and earlier Peak