Multiple comparison correction Methods & models for fMRI data analysis 18 March 2009 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.

Slides:



Advertisements
Similar presentations
The General Linear Model (GLM)
Advertisements

Mkael Symmonds, Bahador Bahrami
Random Field Theory Methods for Dummies 2009 Lea Firmin and Anna Jafarpour.
SPM Software & Resources Wellcome Trust Centre for Neuroimaging University College London SPM Course London, May 2011.
SPM Software & Resources Wellcome Trust Centre for Neuroimaging University College London SPM Course London, October 2008.
Event-related fMRI (er-fMRI)
Multiple comparison correction
Statistical Inference and Random Field Theory Will Penny SPM short course, London, May 2003 Will Penny SPM short course, London, May 2003 M.Brett et al.
Experimental design of fMRI studies Methods & models for fMRI data analysis in neuroeconomics April 2010 Klaas Enno Stephan Laboratory for Social and Neural.
Group analyses of fMRI data Methods & models for fMRI data analysis in neuroeconomics November 2010 Klaas Enno Stephan Laboratory for Social and Neural.
Multiple comparison correction
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research University of Zurich With many thanks for slides & images to: FIL Methods.
Topological Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course London, May 2014 Many thanks to Justin.
Classical inference and design efficiency Zurich SPM Course 2014
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich With.
07/01/15 MfD 2014 Xin You Tai & Misun Kim
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich With.
The General Linear Model (GLM)
Comparison of Parametric and Nonparametric Thresholding Methods for Small Group Analyses Thomas Nichols & Satoru Hayasaka Department of Biostatistics U.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK Institute of Empirical Research in Economics, Zurich, Switzerland Bayesian inference.
Group analyses of fMRI data Methods & models for fMRI data analysis 28 April 2009 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
Multiple comparison correction Methods & models for fMRI data analysis 29 October 2008 Klaas Enno Stephan Branco Weiss Laboratory (BWL) Institute for Empirical.
Group analyses of fMRI data Methods & models for fMRI data analysis 26 November 2008 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
False Discovery Rate Methods for Functional Neuroimaging Thomas Nichols Department of Biostatistics University of Michigan.
Giles Story Philipp Schwartenbeck
2nd Level Analysis Jennifer Marchant & Tessa Dekker
Multiple Comparison Correction in SPMs Will Penny SPM short course, Zurich, Feb 2008 Will Penny SPM short course, Zurich, Feb 2008.
With many thanks for slides & images to: FIL Methods group, Virginia Flanagin and Klaas Enno Stephan Dr. Frederike Petzschner Translational Neuromodeling.
Random Field Theory Will Penny SPM short course, London, May 2005 Will Penny SPM short course, London, May 2005 David Carmichael MfD 2006 David Carmichael.
Basics of fMRI Inference Douglas N. Greve. Overview Inference False Positives and False Negatives Problem of Multiple Comparisons Bonferroni Correction.
Random field theory Rumana Chowdhury and Nagako Murase Methods for Dummies November 2010.
Computational Biology Jianfeng Feng Warwick University.
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich With.
SPM Course Zurich, February 2015 Group Analyses Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London With many thanks to.
Multiple comparison correction Methods & models for fMRI data analysis October 2013 With many thanks for slides & images to: FIL Methods group & Tom Nichols.
Multiple comparisons in M/EEG analysis Gareth Barnes Wellcome Trust Centre for Neuroimaging University College London SPM M/EEG Course London, May 2013.
Group analyses of fMRI data Methods & models for fMRI data analysis November 2012 With many thanks for slides & images to: FIL Methods group, particularly.
1 Inference on SPMs: Random Field Theory & Alternatives Thomas Nichols, Ph.D. Department of Statistics & Warwick Manufacturing Group University of Warwick.
Methods for Dummies Random Field Theory Annika Lübbert & Marian Schneider.
Classical Inference on SPMs Justin Chumbley SPM Course Oct 23, 2008.
Contrasts & Statistical Inference
**please note** Many slides in part 1 are corrupt and have lost images and/or text. Part 2 is fine. Unfortunately, the original is not available, so please.
Random Field Theory Will Penny SPM short course, London, May 2005 Will Penny SPM short course, London, May 2005.
Random Field Theory Ciaran S Hill & Christian Lambert Methods for Dummies 2008.
Event-related fMRI Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course Chicago, Oct 2015.
The False Discovery Rate A New Approach to the Multiple Comparisons Problem Thomas Nichols Department of Biostatistics University of Michigan.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
Statistical Analysis An Introduction to MRI Physics and Analysis Michael Jay Schillaci, PhD Monday, April 7 th, 2007.
FMRI Modelling & Statistical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course Chicago, Oct.
SPM short – Mai 2008 Linear Models and Contrasts Stefan Kiebel Wellcome Trust Centre for Neuroimaging.
Multiple comparisons problem and solutions James M. Kilner
Topological Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM Course London, May 2015 With thanks to Justin.
Multiple comparison correction
False Discovery Rate for Functional Neuroimaging Thomas Nichols Department of Biostatistics University of Michigan Christopher Genovese & Nicole Lazar.
Topological Inference
The general linear model and Statistical Parametric Mapping
2nd Level Analysis Methods for Dummies 2010/11 - 2nd Feb 2011
Wellcome Trust Centre for Neuroimaging University College London
Methods for Dummies Random Field Theory
Multiple comparisons in M/EEG analysis
Topological Inference
Contrasts & Statistical Inference
Inference on SPMs: Random Field Theory & Alternatives
Statistical Parametric Mapping
The general linear model and Statistical Parametric Mapping
Contrasts & Statistical Inference
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
Multiple testing Justin Chumbley Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich.
The General Linear Model (GLM)
Contrasts & Statistical Inference
Presentation transcript:

Multiple comparison correction Methods & models for fMRI data analysis 18 March 2009 Klaas Enno Stephan Laboratory for Social and Neural Systems Research Institute for Empirical Research in Economics University of Zurich Functional Imaging Laboratory (FIL) Wellcome Trust Centre for Neuroimaging University College London With many thanks for slides & images to: FIL Methods group

Overview of SPM RealignmentSmoothing Normalisation General linear model Statistical parametric map (SPM) Image time-series Parameter estimates Design matrix Template Kernel Gaussian field theory p <0.05 Statisticalinference

Time BOLD signal Time single voxel time series single voxel time series Voxel-wise time series analysis model specification model specification parameter estimation parameter estimation hypothesis statistic SPM

Inference at a single voxel  = p(t > u | H 0 ) NULL hypothesis H 0 : activation is zero p-value: probability of getting a value of t at least as extreme as u. If  is small we reject the null hypothesis. We can choose u to ensure a voxel- wise significance level of . t = contrast of estimated parameters variance estimate t distribution  u

Student's t-distribution t-distribution is an approximation to the normal distribution for small samples For high degrees of freedom (large samples), t approximates Z. S n = sample standard deviation  = population standard deviation

Types of error Actual condition Test result Reject H 0 Fail to reject H 0 H 0 trueH 0 false True negative (TN) True positive (TP) False positive (FP) Type I error  False negative (FN) Type II error β specificity: 1-  = TN / (TN + FP) = proportion of actual negatives which are correctly identified sensitivity (power): 1-  = TP / (TP + FN) = proportion of actual positives which are correctly identified

Assessing SPMs t > 0.5 t > 3.5t > 5.5 High ThresholdMed. ThresholdLow Threshold Good Specificity Poor Power (risk of false negatives) Poor Specificity (risk of false positives) Good Power

Inference on images Signal+Noise Noise

Using an ‘uncorrected’ p-value of 0.1 will lead us to conclude on average that 10% of voxels are active when they are not. This is clearly undesirable. To correct for this we can define a null hypothesis for images of statistics.

Family-wise null hypothesis FAMILY-WISE NULL HYPOTHESIS: Activation is zero everywhere. If we reject a voxel null hypothesis at any voxel, we reject the family-wise null hypothesis A false-positive anywhere in the image gives a Family Wise Error (FWE). Family-Wise Error (FWE) rate = ‘corrected’ p-value

Use of ‘uncorrected’ p-value,  =0.1 FWE Use of ‘corrected’ p-value,  =0.1

The Bonferroni correction The family-wise error rate (FWE), ,a family of N independent The family-wise error rate (FWE), , for a family of N independent voxels is α = Nv where v is the voxel-wise error rate. Therefore, to ensure a particular FWE, we can use v = α / N BUT...

The Bonferroni correction Independent voxelsSpatially correlated voxels Bonferroni correction assumes independence of voxels  this is too conservative for smooth brain images !

Smoothness (or roughness) roughness = 1/smoothness intrinsic smoothness –some vascular effects have extended spatial support extrinsic smoothness –resampling during preprocessing –matched filter theorem  deliberate additional smoothing to increase SNR described in resolution elements: "resels" resel = size of image part that corresponds to the FWHM (full width half maximum) of the Gaussian convolution kernel that would have produced the observed image when applied to independent voxel values # resels is similar, but not identical to # independent observations can be computed from spatial derivatives of the residuals

Random Field Theory Consider a statistic image as a discretisation of a continuous underlying random field with a certain smoothness Use results from continuous random field theory Discretisation (“lattice approximation”)

Euler characteristic (EC) Topological measure –threshold an image at u -EC = # blobs -at high u: p (blob) = E [EC] therefore (under H 0 ) FWE,  = E [EC]

Euler characteristic (EC) for 2D images R= number of resels Z T = Z value threshold We can determine that Z threshold for which E[EC] = At this threshold, every remaining voxel represents a significant activation, corrected for multiple comparisons across the search volume. Example: For 100 resels, E [EC] = for a Z threshold of 3.8. That is, the probability of getting one or more blobs where Z is greater than 3.8, is Expected EC values for an image of 100 resels

Euler characteristic (EC) for any image Computation of E[EC] can be generalized to be valid for volumes of any dimensions, shape and size, including small volumes (Worsley et al. 1996, A unified statistical approach for determining significant signals in images of cerebral activation, Human Brain Mapping, 4, 58–83.) When we have a good a priori hypothesis about where an activation should be, we can reduce the search volume: –mask defined by (probabilistic) anatomical atlases –mask defined by separate "functional localisers" –mask defined by orthogonal contrasts –spherical search volume around known coordinates small volume correction (SVC)

Voxel level test: intensity of a voxel Cluster level test: spatial extent above u Set level test: number of clusters above u Sensitivity  Regional specificity  Voxel, cluster and set level tests

False Discovery Rate (FDR) Familywise Error Rate (FWE) –probability of one or more false positive voxels in the entire image False Discovery Rate (FDR) –FDR = E(V/R) (R voxels declared active, V falsely so) –proportion of activated voxels that are false positives

False Discovery Rate - Illustration Signal Signal+Noise Noise

FWE 6.7% 10.4%14.9%9.3%16.2%13.8%14.0% 10.5%12.2%8.7% Control of Familywise Error Rate at 10% 11.3% 12.5%10.8%11.5%10.0%10.7%11.2%10.2%9.5% Control of Per Comparison Rate at 10% Percentage of False Positives Control of False Discovery Rate at 10% Occurrence of Familywise Error Percentage of Activated Voxels that are False Positives

Benjamini & Hochberg procedure Select desired limit q on FDR Order p-values, p (1)  p (2) ...  p (V) Let r be largest i such that Reject all null hypotheses corresponding to p (1),..., p (r). p (i)  (i/V)  q p(i)p(i) i/Vi/V (i/V)  q(i/V)  q p-value Benjamini & Hochberg, JRSS-B (1995) 57: i/V = proportion of all selected voxels

Real Data: FWE correction with RFT Threshold –S = 110,776 –2  2  2 voxels 5.1  5.8  6.9 mm FWHM –u = Result –5 voxels above the threshold -log 10 p-value

Threshold –u = 3.83 Result –3,073 voxels above threshold Real Data: FWE correction with FDR

Caveats concerning FDR Current methodological discussions whether standard FDR implementations are valid for neuroimaging data Some argue (Chumbley & Friston 2009, NeuroImage) that the fMRI signal is spatially extended, it does not have compact support → inference should therefore not be about single voxels, but about topological features of the signal (e.g. peaks or clusters) In contrast, FDR=E(V/R), i.e. the expected fraction of all positive decisions R, that are false positive decisions V. To be applicable, this definition requires that a subset of the image is signal-free. In images with continuous signal (e.g. after smoothing), all voxels have signal and consequently there are no false positives; FDR (and FWE) must be zero. Possible alternative: FDR on topological features (e.g. clusters)

Conclusions Corrections for multiple testing are necessary to control the false positive risk. FWE –Very specific, not so sensitive –Random Field Theory Inference about topological features (peaks, clusters) Excellent for large sample sizes (e.g. single-subject analyses or large group analyses) Afford littles power for group studies with small sample size  consider non-parametric methods (not discussed in this talk) FDR –Less specific, more sensitive –Interpret with care! represents false positive risk over whole set of selected voxels voxel-wise inference (which has been criticised)

Thank you