Statistical Parametric Mapping Will Penny Wellcome Trust Centre

  • Slides: 37
Download presentation
Statistical Parametric Mapping Will Penny Wellcome Trust Centre for Neuroimaging, University College London, UK

Statistical Parametric Mapping Will Penny Wellcome Trust Centre for Neuroimaging, University College London, UK LSTHM, UCL, Jan 14, 2009

Statistical Parametric Mapping Image time-series Kernel Design matrix Realignment Smoothing General linear model Statistical

Statistical Parametric Mapping Image time-series Kernel Design matrix Realignment Smoothing General linear model Statistical Parametric Map Statistical inference Normalisation Random field theory p <0. 05 Template Parameter estimates

Outline • Voxel-wise General Linear Models • Random Field Theory • Bayesian modelling

Outline • Voxel-wise General Linear Models • Random Field Theory • Bayesian modelling

Voxel-wise GLMs model specification Time parameter estimation hypothesis statistic Ti e m Intensity single

Voxel-wise GLMs model specification Time parameter estimation hypothesis statistic Ti e m Intensity single voxel time series SPM

Temporal convolution model for the BOLD response Convolve stimulus function with a canonical hemodynamic

Temporal convolution model for the BOLD response Convolve stimulus function with a canonical hemodynamic response function (HRF): HRF

General Linear Model = + Error Covariance N: number of scans p: number of

General Linear Model = + Error Covariance N: number of scans p: number of regressors Model is specified by 1. Design matrix X 2. Assumptions about e

Estimation 1. Re. ML-algorithm L g l 2. Weighted Least Squares Friston et al.

Estimation 1. Re. ML-algorithm L g l 2. Weighted Least Squares Friston et al. 2002, Neuroimage

Contrasts & SPMs c=100000 Q: activation during listening ? Null hypothesis:

Contrasts & SPMs c=100000 Q: activation during listening ? Null hypothesis:

Outline • Voxel-wise General Linear Models • Random Field Theory • Bayesian modelling

Outline • Voxel-wise General Linear Models • Random Field Theory • Bayesian modelling

Inference for Images Noise Signal+Noise

Inference for Images Noise Signal+Noise

Use of ‘uncorrected’ p-value, a=0. 1 11. 3% 12. 5% 10. 8% 11. 5%

Use of ‘uncorrected’ p-value, a=0. 1 11. 3% 12. 5% 10. 8% 11. 5% 10. 0% 10. 7% 11. 2% 10. 2% 9. 5% Percentage of Null Pixels that are False Positives Using an ‘uncorrected’ p-value of 0. 1 will lead us to conclude on average that 10% of voxels are active when they are not. This is clearly undesirable. To correct for this we can define a null hypothesis for images of statistics.

Family-wise Null Hypothesis FAMILY-WISE NULL HYPOTHESIS: Activation is zero everywhere If we reject a

Family-wise Null Hypothesis FAMILY-WISE NULL HYPOTHESIS: Activation is zero everywhere If we reject a voxel null hypothesis at any voxel, we reject the family-wise Null hypothesis A FP anywhere in the image gives a Family Wise Error (FWE) Family-Wise Error (FWE) rate = ‘corrected’ p-value

Use of ‘uncorrected’ p-value, a=0. 1 Use of ‘corrected’ p-value, a=0. 1 FWE

Use of ‘uncorrected’ p-value, a=0. 1 Use of ‘corrected’ p-value, a=0. 1 FWE

Spatial correlation Independent Voxels Spatially Correlated Voxels

Spatial correlation Independent Voxels Spatially Correlated Voxels

Random Field Theory • Consider a statistic image as a discretisation of a continuous

Random Field Theory • Consider a statistic image as a discretisation of a continuous underlying random field • Use results from continuous random field theory Discretisation

Euler Characteristic (EC) Topological measure – threshold an image at u - EC =

Euler Characteristic (EC) Topological measure – threshold an image at u - EC = # blobs - at high u: Prob blob = avg (EC) so FWE, a = avg (EC)

Example – 2 D Gaussian images α = R (4 ln 2) (2π) -3/2

Example – 2 D Gaussian images α = R (4 ln 2) (2π) -3/2 u exp (-u 2/2) Voxel-wise threshold, u Number of Resolution Elements (RESELS), R N=100 x 100 voxels, Smoothness FWHM=10, gives R=10 x 10=100

Example – 2 D Gaussian images α = R (4 ln 2) (2π) -3/2

Example – 2 D Gaussian images α = R (4 ln 2) (2π) -3/2 u exp (-u 2/2) For R=100 and α=0. 05 RFT gives u=3. 8

SPM results

SPM results

SPM results. . .

SPM results. . .

Outline • Voxel-wise General Linear Models • Random Field Theory • Bayesian Modelling

Outline • Voxel-wise General Linear Models • Random Field Theory • Bayesian Modelling

Motivation Even without applied spatial smoothing, activation maps (and maps of eg. AR coefficients)

Motivation Even without applied spatial smoothing, activation maps (and maps of eg. AR coefficients) have spatial structure Contrast AR(1) We can increase the sensitivity of our inferences by smoothing data with Gaussian kernels (SPM 2). This is worthwhile, but crude. Can we do better with a spatial model (SPM 5) ? Aim: For SPM 5 to remove the need for spatial smoothing just as SPM 2 removed the need for temporal smoothing

The Model u 1 q 2 r 1 u 2 a b l W

The Model u 1 q 2 r 1 u 2 a b l W A Voxel-wise AR: Y Y=XW+E [Tx. N] [Tx. K] [Kx. N] [Tx. N] Spatial pooled AR:

Synthetic Data 1 : from Laplacian Prior reshape(w 1, 32) t

Synthetic Data 1 : from Laplacian Prior reshape(w 1, 32) t

Prior, Likelihood and Posterior In the prior, W factorises over k and A factorises

Prior, Likelihood and Posterior In the prior, W factorises over k and A factorises over p: The likelihood factorises over n: The posterior over W therefore does’nt factor over k or n. It is a Gaussian with an NK-by-NK full covariance matrix. This is unwieldy to even store, let alone invert ! So exact inference is intractable.

Variational Bayes L KL F

Variational Bayes L KL F

Variational Bayes If you assume posterior factorises then F can be maximised by letting

Variational Bayes If you assume posterior factorises then F can be maximised by letting where

Variational Bayes In the prior, W factorises over k and A factorises over p:

Variational Bayes In the prior, W factorises over k and A factorises over p: In chosen approximate posterior, W and A factorise over n: So, in the posterior for W we only have to store and invert N K-by-K covariance matrices.

Updating approximate posterior Regression coefficients, W Spatial precisions for W AR coefficients, A Spatial

Updating approximate posterior Regression coefficients, W Spatial precisions for W AR coefficients, A Spatial precisions for A Observation noise

Synthetic Data 1 : from Laplacian Prior y x t

Synthetic Data 1 : from Laplacian Prior y x t

F Iteration Number

F Iteration Number

VB – Laplacian Prior Least Squares y y x Coefficients = 1024 x `Coefficient

VB – Laplacian Prior Least Squares y y x Coefficients = 1024 x `Coefficient RESELS’ = 366

Synthetic Data II : blobs True Global prior Smoothing Laplacian prior

Synthetic Data II : blobs True Global prior Smoothing Laplacian prior

Sensitivity 1 -Specificity

Sensitivity 1 -Specificity

Event-related f. MRI: Faces versus chequerboard Smoothing Global prior Laplacian Prior

Event-related f. MRI: Faces versus chequerboard Smoothing Global prior Laplacian Prior

Event-related f. MRI: Familiar faces versus unfamiliar faces Smoothing Penny WD, Trujillo. Barreto NJ,

Event-related f. MRI: Familiar faces versus unfamiliar faces Smoothing Penny WD, Trujillo. Barreto NJ, Friston KJ. Bayesian f. MRI time series analysis with spatial priors. Neuroimage. 2005 Jan 15; 24(2): 350 -62. Global prior Laplacian Prior

Summary • Voxel-wise General Linear Models • Random Field Theory • Bayesian Modelling http:

Summary • Voxel-wise General Linear Models • Random Field Theory • Bayesian Modelling http: //www. fil. ion. ucl. ac. uk/~wpenny/mbi/index. html Graph-partitioned spatial priors for functional magnetic resonance images. Harrison LM, Penny W, Flandin G, Ruff CC, Weiskopf N, Friston KJ. Neuroimage. 2008 Dec; 43(4): 694 -707.