Transcript DCM - UZH
The General Linear Model (GLM) Klaas Enno Stephan Branco Weiss Laboratory (BWL) Institute for Empirical Research in Economics University of Zurich With many thanks for slides & images to: FIL Methods group Functional Imaging Laboratory (FIL) Wellcome Trust Centre for Neuroimaging University College London Methods & models for fMRI data analysis 15 October 2008 Overview of SPM Image time-series Realignment Kernel Design matrix Smoothing General linear model Statistical parametric map (SPM) Statistical inference Normalisation Gaussian field theory p <0.05 Template Parameter estimates A very simple fMRI experiment One session Passive word listening versus rest 7 cycles of rest and listening Blocks of 6 scans with 7 sec TR Stimulus function Question: Is there a change in the BOLD response between listening and rest? Modelling the measured data Why? How? Make inferences about effects of interest 1. Decompose data into effects and error 2. Form statistic using estimates of effects and error stimulus function data linear model effects estimate error estimate statistic Voxel-wise time series analysis model specification Time parameter estimation hypothesis statistic BOLD signal single voxel time series SPM Time =1 BOLD signal + 2 x1 + x2 y x11 x2 2 e error Single voxel regression model e Mass-univariate analysis: voxel-wise GLM p 1 1 1 p y N = N X y X e e ~ N (0, I ) 2 + N e Model is specified by 1. Design matrix X 2. Assumptions about e N: number of scans p: number of regressors The design matrix embodies all available knowledge about experimentally controlled factors and potential confounds. GLM assumes Gaussian “spherical” (i.i.d.) errors sphericity = iid: error covariance is scalar multiple of identity matrix: Cov(e) = 2I Examples for non-sphericity: 4 0 Cov(e) 0 1 non-identity 1 0 Cov(e) 0 1 2 1 Cov(e) 1 2 non-identity non-independence Parameter estimation 1 + 2 = y Objective: estimate parameters to minimize X y X e e N e t 1 Ordinary least squares estimation (OLS) (assuming i.i.d. error): ˆ ( X T X ) 1 X T y 2 t A geometric perspective on the GLM Residual forming matrix R OLS estimates y T 1 T ˆ (X X ) X y e yˆ Xˆ x2 x1 Design space defined by X e Ry R I P Projection matrix P yˆ Py 1 P X (X X ) X T T Correlated and orthogonal regressors y x2* x2 x1 y x11 x2 2 e y x11 x2* 2* e 1 2 1 1 1; 2* 1 Correlated regressors = explained variance is shared between regressors When x2 is orthogonalized with regard to x1, only the parameter estimate for x1 changes, not that for x2! What are the problems of this model? 1. BOLD responses have a delayed and dispersed form. 2. The BOLD signal includes substantial amounts of lowfrequency noise. 3. The data are serially correlated (temporally autocorrelated) this violates the assumptions of the noise model in the GLM HRF Problem 1: Shape of BOLD response Solution: Convolution model hemodynamic response function (HRF) t f g (t ) f ( ) g (t )d 0 The response of a linear time-invariant (LTI) system is the convolution of the input with the system's response to an impulse (delta function). expected BOLD response = input function impulse response function (HRF) Convolution model of the BOLD response Convolve stimulus function with a canonical hemodynamic response function (HRF): t f g (t ) f ( ) g (t )d 0 HRF Problem 2: Low-frequency noise Solution: High pass filtering Sy SX Se S = residual forming matrix of DCT set discrete cosine transform (DCT) set High pass filtering: example blue = data black = mean + low-frequency drift green = predicted response, taking into account low-frequency drift red = predicted response, NOT taking into account low-frequency drift Problem 3: Serial correlations et aet 1 t with t ~ N (0, 2 ) 1st order autoregressive process: AR(1) N Cov(e) autocovariance function N Dealing with serial correlations • Pre-colouring: impose some known autocorrelation structure on the data (filtering with matrix W) and use Satterthwaite correction for df’s. • Pre-whitening: 1. Use an enhanced noise model with multiple error covariance components. 2. Use estimated autocorrelation to specify filter matrix W for whitening the data. Wy WX We How do we define V ? • Enhanced noise model • Remember linear transform for Gaussians • Choose W such that error covariance becomes spherical e ~ N (0, V ) 2 x ~ N ( , ), y ax 2 y ~ N (a , a 2 2 ) We ~ N (0, W V ) 2 W V I 2 • Conclusion: W is a function of V so how do we estimate V ? W V Wy WX We 1 / 2 2 Multiple covariance components V Cov(e) e ~ N (0, V ) 2 enhanced noise model V = 1 V iQi error covariance components Q and hyperparameters Q1 + 2 Q2 Estimation of hyperparameters with ReML (restricted maximum likelihood). Contrasts & statistical parametric maps c=10000000000 Q: activation during listening ? X Null hypothesis: 1 0 c ˆ t T ˆ Std (c ) T t-statistic based on ML estimates Wy WX We ˆ (WX ) Wy c=10000000000 c ˆ t stˆd (cT ˆ ) T W V stˆd (cT ˆ ) ˆ c (WX ) (WX ) c 1/ 2 ˆ 2 V Cov(e) 2 2 T T Wy WXˆ tr( R) R I WX (WX ) X V Q i ReMLestimates i 2 Physiological confounds • head movements • arterial pulsations • breathing • eye blinks • adaptation affects, fatigue, fluctuations in concentration, etc. Outlook: further challenges • correction for multiple comparisons • variability in the HRF across voxels • slice timing • limitations of frequentist statistics Bayesian analyses • GLM ignores interactions among voxels models of effective connectivity These issues are discussed in future lectures. Correction for multiple comparisons • Mass-univariate approach: We apply the GLM to each of a huge number of voxels (usually > 100,000). • Threshold of p<0.05 more than 5000 voxels significant by chance! • Massive problem with multiple comparisons! • Solution: Gaussian random field theory Variability in the HRF • HRF varies substantially across voxels and subjects • For example, latency can differ by ± 1 second • Solution: use multiple basis functions • See talk on event-related fMRI Summary • Mass-univariate approach: same GLM for each voxel • GLM includes all known experimental effects and confounds • Convolution with a canonical HRF • High-pass filtering to account for low-frequency drifts • Estimation of multiple variance components (e.g. to account for serial correlations) • Parametric statistics