1 / 60

Zurich SPM Course 2014 14 February 2014

DCM: Advanced topics. Klaas Enno Stephan . Zurich SPM Course 2014 14 February 2014. Overview. Generative models & analysis options Extended DCM for fMRI : nonlinear , two-state , stochastic Embedding computational models into DCMs Integrating tractography and DCM

garret
Télécharger la présentation

Zurich SPM Course 2014 14 February 2014

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. DCM: Advanced topics Klaas Enno Stephan Zurich SPM Course 2014 14 February 2014

  2. Overview • Generative models & analysisoptions • Extended DCM forfMRI: nonlinear, two-state, stochastic • Embedding computational models into DCMs • Integratingtractographyand DCM • ApplicationsofDCM toclinicalquestions

  3. Generative models Advantages: forceustothinkmechanistically: howwerethedatacaused? allowonetogeneratesyntheticdata. Bayesianperspective→ inversion & modelevidence

  4. Dynamic Causal Modeling (DCM) Hemodynamicforward model:neural activityBOLD Electromagnetic forward model:neural activityEEGMEG LFP Neural state equation: fMRI EEG/MEG simple neuronal model complicated forward model complicated neuronal model simple forward model inputs

  5. Bayesian system identification Design experimental inputs Neural dynamics Define likelihood model Observer function Specify priors Inference on model structure Invert model Inference on parameters Make inferences

  6. VB in a nutshell (mean-field approximation)  Neg. free-energy approx. to model evidence.  Mean field approx.  Maximise neg. free energy wrt. q = minimise divergence, by maximisingvariational energies  Iterative updating of sufficient statistics of approx. posteriors by gradient ascent.

  7. Generative models • any DCM = a particular generative model of how the data (may) have been caused • modelling = comparing competing hypotheses about the mechanisms underlyingobserveddata • modelspace:a priori definitionofhypothesissetiscrucial • modelselection:determinethemost plausible hypothesis (model), giventhedata • inference on parameters: e.g., evaluateconsistencyofhowmodelmechanismsareimplementedacrosssubjects • model selection  model validation! • model validation requires external criteria (external to the measured data)

  8. Pitt & Miyung (2002) TICS Model comparison and selection Given competing hypotheses on structure & functional mechanisms of a system, which model is the best? Which model represents thebest balance between model fit and model complexity? For which model m does p(y|m) become maximal?

  9. Bayesian model selection (BMS) Model evidence: Gharamani, 2004 p(y|m) y all possible datasets accounts for both accuracy and complexity of the model • Various approximations, e.g.: • negative free energy, AIC, BIC a measure of generalizability McKay 1992, Neural Comput. Penny et al. 2004a, NeuroImage

  10. Approximations to the model evidence in DCM Maximizing log model evidence = Maximizing model evidence Logarithm is a monotonic function Log model evidence = balance between fit and complexity No. of parameters In SPM2 & SPM5, interface offers 2 approximations: No. of data points Akaike Information Criterion: Bayesian Information Criterion: Penny et al. 2004a, NeuroImage

  11. The (negative) free energy approximation • UnderGaussianassumptionsabouttheposterior (Laplace approximation):

  12. The complexity term in F • In contrastto AIC & BIC, thecomplexitytermofthe negative freeenergyFaccountsforparameterinterdependencies. • The complexitytermofFishigher • themoreindependentthepriorparameters ( effective DFs) • themoredependenttheposteriorparameters • themoretheposteriormeandeviatesfromthepriormean • NB: Since SPM8, onlyFisusedformodelselection !

  13. definition of model space inference on model structure or inference on model parameters? inference on individual models or model space partition? inference on parameters of an optimal model or parameters of all models? optimal model structure assumed to be identical across subjects? comparison of model families using FFX or RFX BMS optimal model structure assumed to be identical across subjects? BMA yes no yes no FFX BMS RFX BMS FFX BMS RFX BMS FFX analysis of parameter estimates (e.g. BPA) RFX analysis of parameter estimates (e.g. t-test, ANOVA) Stephan et al. 2010, NeuroImage

  14. Random effects BMS forheterogeneousgroups Dirichlet parameters  = “occurrences” of models in the population Dirichlet distribution of model probabilities r Multinomial distribution of model labels m Model inversion by VariationalBayes or MCMC Measured data y Stephan et al. 2009a, NeuroImage Penny et al. 2010, PLoSComput. Biol.

  15. Inference about DCM parameters:Bayesian single-subject analysis • Gaussian assumptions about the posterior distributions of the parameters • Use of the cumulative normal distribution to test the probability that a certain parameter (or contrast of parameters cTηθ|y) is above a chosen threshold γ: • By default, γ is chosen as zero ("does the effect exist?").

  16. Inference about DCM parameters:Random effects group analysis (classical) • In analogy to “random effects” analyses in SPM, 2nd level analyses can be applied to DCM parameters: Separate fitting of identical models for each subject Selection of model parameters of interest one-sample t-test:parameter > 0 ? paired t-test:parameter 1 > parameter 2 ? rmANOVA:e.g. in case of multiple sessions per subject

  17. Bayesian Model Averaging (BMA) • uses the entire model space considered (or an optimal family of models) • averages parameter estimates, weighted by posterior model probabilities • particularly useful alternative when • none of the models (subspaces) considered clearly outperforms all others • when comparing groups for which the optimal model differs NB: p(m|y1..N) can be obtained by either FFX or RFX BMS Penny et al. 2010, PLoS Comput. Biol.

  18. definition of model space inference on model structure or inference on model parameters? inference on individual models or model space partition? inference on parameters of an optimal model or parameters of all models? optimal model structure assumed to be identical across subjects? comparison of model families using FFX or RFX BMS optimal model structure assumed to be identical across subjects? BMA yes no yes no FFX BMS RFX BMS FFX BMS RFX BMS FFX analysis of parameter estimates (e.g. BPA) RFX analysis of parameter estimates (e.g. t-test, ANOVA) Stephan et al. 2010, NeuroImage

  19. Overview • Generative models & analysisoptions • Extended DCM forfMRI: nonlinear, two-state, stochastic • Embedding computational models into DCMs • Integratingtractographyand DCM • ApplicationsofDCM toclinicalquestions

  20. The evolution of DCM in SPM • DCM is not one specific model, but a framework for Bayesian inversion of dynamic system models • The default implementation in SPM is evolving over time • improvementsofnumericalroutines (e.g., forinversion) • change in parameterization (e.g., self-connections, hemodynamicstates in log space) • change in priorstoaccommodatenew variants (e.g., stochastic DCMs, endogenous DCMs etc.) To enable replication of your results, you should ideally state which SPM version (releasenumber) you are using when publishing papers. The releasenumberisstored in theDCM.matfile.

  21. Neural state equation endogenous connectivity modulation of connectivity direct inputs modulatory input u2(t) driving input u1(t) t t y BOLD y y y   λ hemodynamic model  activity x2(t) activity x3(t) activity x1(t) x neuronal states integration The classical DCM: a deterministic, one-state, bilinear model

  22. Factorial structure of model specification in DCM10 • Three dimensions of model specification: • bilinear vs. nonlinear • single-state vs. two-state (per region) • deterministic vs. stochastic • Specification via GUI.

  23. non-linear DCM modulation driving input bilinear DCM driving input modulation Two-dimensional Taylor series (around x0=0, u0=0): Nonlinear state equation: Bilinear state equation:

  24. Neural population activity x3 fMRI signal change (%) x1 x2 u2 u1 Nonlinear dynamic causal model (DCM) Stephan et al. 2008, NeuroImage

  25. attention MAP = 1.25 0.10 PPC 0.26 0.39 1.25 0.26 V1 stim 0.13 V5 0.46 0.50 motion Stephan et al. 2008, NeuroImage

  26. Two-state DCM Single-state DCM Two-state DCM input Extrinsic (between-region) coupling Intrinsic (within-region) coupling Marreiros et al. 2008, NeuroImage

  27. Estimates of hidden causes and states (Generalised filtering) Stochastic DCM • all states are represented in generalised coordinates of motion • random state fluctuations w(x)account for endogenous fluctuations,have unknown precision and smoothness  two hyperparameters • fluctuations w(v) induce uncertainty about how inputs influence neuronal activity • can be fitted to resting state data Li et al. 2011, NeuroImage

  28. Overview • Generative models & analysisoptions • Extended DCM forfMRI: nonlinear, two-state, stochastic • Embedding computational models in DCMs • Integratingtractographyand DCM • Applicationsof DCM toclinicalquestions

  29. Prediction errors drive synaptic plasticity PE(t) x3 R x1 x2 McLaren 1989 synaptic plasticity during learning = f (prediction error)

  30. Conditioning Stimulus Target Stimulus or 1 0.8 or 0.6 CS TS Response 0.4 0 200 400 600 800 2000 ± 650 CS 1 Time (ms) CS 0.2 2 0 0 200 400 600 800 1000 Learning ofdynamic audio-visualassociations p(face) trial den Ouden et al. 2010, J. Neurosci.

  31. k vt-1 vt rt rt+1 ut ut+1 Hierarchical Bayesian learning model prior on volatility volatility probabilistic association observed events Behrens et al. 2007, Nat. Neurosci.

  32. 1 True Bayes Vol HMM fixed 0.8 HMM learn RW 0.6 p(F) 450 0.4 440 0.2 430 RT (ms) 420 0 400 440 480 520 560 600 Trial 410 400 390 0.1 0.3 0.5 0.7 0.9 p(outcome) Explaining RTs by different learning models Reaction times Bayesian model selection: hierarchical Bayesianmodel performsbest • 5 alternative learning models: • categorical probabilities • hierarchical Bayesian learner • Rescorla-Wagner • Hidden Markov models (2 variants) den Ouden et al. 2010, J. Neurosci.

  33. p < 0.05 (SVC) 0 0 -0.5 -0.5 BOLD resp. (a.u.) BOLD resp. (a.u.) -1 -1 -1.5 -1.5 -2 -2 p(F) p(H) p(F) p(H) Stimulus-independent prediction error Putamen Premotor cortex p < 0.05 (cluster-level whole- brain corrected) den Ouden et al. 2010, J. Neurosci.

  34. Prediction error (PE) activity in the putamen PE duringactive sensorylearning PE duringincidental sensorylearning den Ouden et al. 2009, Cerebral Cortex p < 0.05 (SVC) PE during reinforcement learning PE = “teaching signal” for synaptic plasticity during learning O'Doherty et al. 2004, Science Could the putamen be regulating trial-by-trial changes of task-relevant connections?

  35. Prediction errors control plasticity during adaptive cognition Hierarchical Bayesian learning model PUT • Influence of visual areas on premotor cortex: • stronger for surprising stimuli • weaker for expected stimuli p= 0.017 p= 0.010 PMd PPA FFA den Ouden et al. 2010, J. Neurosci.

  36. Overview • Generative models & analysisoptions • Extended DCM forfMRI: nonlinear, two-state, stochastic • Embedding computational models in DCMs • Integratingtractographyand DCM • Applicationsof DCM toclinicalquestions

  37. Diffusion-weighted imaging Parker & Alexander, 2005, Phil. Trans. B

  38. Probabilistic tractography: Kaden et al. 2007, NeuroImage • computes local fibre orientation density by spherical deconvolution of the diffusion-weighted signal • estimates the spatial probability distribution of connectivity from given seed regions • anatomical connectivity = proportion of fibre pathways originating in a specific source region that intersect a target region • If the area or volume of the source region approaches a point, this measure reduces to method by Behrens et al. (2003)

  39. Integration of tractography and DCM R1 R2 low probability of anatomical connection  small prior variance of effective connectivity parameter R1 R2 high probability of anatomical connection  large prior variance of effective connectivity parameter Stephan, Tittgemeyer et al. 2009, NeuroImage

  40. probabilistic tractography FG right LG right anatomicalconnectivity LG left FG left LG LG FG FG Proofofconceptstudy DCM connection-specificpriorsforcouplingparameters Stephan, Tittgemeyer et al. 2009, NeuroImage

  41. Connection-specific prior variance  as a function of anatomical connection probability  • 64 different mappings by systematic search across hyper-parameters  and  • yields anatomically informed (intuitive and counterintuitive) and uninformed priors

  42. Models with anatomically informed priors (of an intuitive form)

  43. Models with anatomically informed priors (of an intuitive form) were clearly superiortoanatomically uninformed ones: Bayes Factor >109

  44. Overview • Generative models & analysisoptions • Extended DCM for fMRI: nonlinear, two-state, stochastic • Embedding computational models in DCMs • Integrating tractography andDCM • Applicationsof DCM toclinicalquestions

  45. Model-based predictions for single patients model structure BMS set of parameter estimates model-based decoding

  46. BMS: Parkison‘s disease and treatment Age-matched controls PD patients on medication PD patients off medication Selection of action modulates connections between PFC and SMA DA-dependent functional disconnection of the SMA Rowe et al. 2010, NeuroImage

  47. Model-based decoding by generative embedding A A step 1 — model inversion step 2 — kernel construction A → B A → C B → B B → C B B C C measurements from an individual subject subject-specificinverted generative model subject representation in the generative score space step 3 — support vector classification step 4 — interpretation jointly discriminative model parameters separating hyperplane fitted to discriminate between groups Brodersen et al. 2011, PLoS Comput. Biol.

  48. Discovering remote or “hidden” brain lesions

  49. Discovering remote or “hidden” brain lesions

  50. Model-based decoding of disease status: mildly aphasic patients (N=11) vs. controls (N=26) Connectional fingerprints from a 6-region DCM of auditory areas during speech perception Brodersen et al. 2011, PLoS Comput. Biol.

More Related