The General Linear Model and Single-Subject & Group Analyses Flashcards

1
Q

When are the data ready for statistical analysis?

A

After reconstruction, realignment, stereotactic normalisation and possibly smoothing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the two steps for statistical analysis?

A
  1. Statistics indicating evidence against a null hypothesis of no effect at each voxel are computed - this results in an image of statistics
  2. The statistic image must be assessed, reliably locating voxels where an effect is exhibited whilst limiting the possibility of false positives
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is used for analysis of functional mapping experiments?

A

SPM

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is auditory block-design experiment?

A

• One session
- One subject in the scanner
• Passive word listening versus rest
• 7 cycles of rest and listening
• Blocks of 6 scans with 7 second TR
• In the data template in the brain – what are the involved time series that correspond to the stimulus function
• There is greater response in the brain when you are listening to words vs resting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Why do we model the measured data?

A

Make inferences about effects of interest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How is the measured data modelled?

A

Decompose data into effects and error

Form statistic using estimates of effects and error

Partition the data into 2 points: noise and signal

Get data partition data using linear model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Why is statistic done?

A

To find the ratio between effects estimate divided by error estimate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

When do you get a significant effect?

A

Very small effect size but a very consistent one

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What happens after stimulus function?

A
  1. Data
  2. Linear model
  3. Effect estimate
    - Error estimate
  4. Statistic
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

(Mass univariate) voxel-wise time series analysis:

A

• As the data has been corrected for movement and normalised – look at the single voxel over time
• It corresponds to 1 time series – one time series per voxel
• Do modelling of the time series and describe how we expect it to evolve over time
- Model specification
- Parameter estimation
- Hypothesis
- Statistic
• Repeat the same procedure for each and every voxel in the brain
• We use the same voxel-wise analysis for all the voxel in the brain one at a time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is a single voxel regression model?

A

• If I take a single voxel – volume density over time at a given voxel
• Express it as a linear combination of two things
- Stimulus function e.g. listening to words and resting [e.g. rest = 0, 1= listening]
- Regressor that is constant – it is one for every time point – model the average point of the time series
• Data that is observed can be expressed in a linear combination of 2 terms B1 + B2 [weights are given – unknown parameters]
• Error – noise that is found on top of our measurements
• Data y [time series] is expressed by linear combinations of B1 + B2 of 2 vectors – stimulus function and the average signal and plus error term  linear model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the equation for the single voxel regression model?

A

Y = x1B1 + X2B2 + e

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the General Linear Model (GLM) described by?

A

data Y – 1 x N data points expressed as a linear combination of X and B
• X contains all the columns – everything that you expect to see in the data
• B is the vectors of the unknown parameters – unknown quantities that we want to estimate – how much of the response is observed in a particular voxel
• Error term is the time series the same dimension as the data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the GLM model specified by?

A
  1. Design matrix X

2. Assumptions about e

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How is the time series expressed as?

A

an image – dark regions mean small values and light regions means higher values  FMRI time series in voxels

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the time series expressed as?

A

Linear combination of stimulus function and the constant term + error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

\why does the parameters need to be estimated?

A

Minimize the data - get the data to be as close as possible to 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What are problems of this model with fMRI time series?

A
  1. The BOLD response has a delayed and dispersed shape
  2. The BOLD signal includes substantial amount of low-frequency noise (e.g. due to scanner drift)
  3. Due to breathing, heartbeat and unmodelled neuronal activity, the errors are serially correlated. This violates the assumptions of the noise model in the glm
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

The BOLD response has a delayed and dispersed shape

A
  • The neuronal activity that occur in the brain will be the same thing as measured in FMRI data
  • E.g. if you have a burst of activity at T=0
  • The Haemodynamic response function (HRF) – the response to burst activity at time T=0
  • Take neuronal activity and convolute with HRF to get an Expected BOLD
20
Q

The BOLD signal includes substantial amounts of low-frequency noise (e.g. due to scanner drift

A
  • At the start of the acquisition the mean data will be getting lower and lower because of heating and that has a direct effect on the intensity of the signal  slow decay over time
  • Implement a model by giving a number of regressors that correspond to a cosine function of different frequencies that allows us to model anything up to a certain frequency
  • Get the matrix and fit it to the data
  • Solution is high pass filtering
21
Q

Due to breathing, heartbeat and unmodelled neuronal activity, the errors are serially correlated. This violates the assumptions of the noise model in the GLM

A
  • Noise is i.i.d = independent, identical, distributed

- The model has to take into account the correlation in time

22
Q

What are contrasts?

A

Compare different conditions
• Contrasts are a way to quantify the parameter you are really interested in e.g. difference in response between 2 different conditions
• A contrast c is a vector of length p
• Compute the linear combination of this contrast in your B – unknown parameters
• A contrast selects a specific effect of interest

23
Q

What is T-test?

A

• Passive word listening vs rest
• Compute the test statistics
• Q = activation during listening?
• Null hypothesis: B1=0
• How to specify B1=0
- Contrast of 1 followed by 0
• Compute test statistics equal to the effect size of interest B1 divided by the standard error
• Compute T for every voxel in the brain – get images of test statistics
• Get bilateral activity in the cortex corresponding to the task

24
Q

What is inference at a single voxel?

A
  • Null hypothesis H0 assumes that there is no activation
  • If you compute a test statistic – you assume that you know the distribution of the statistic test – most of the time it will be 0
  • It will be a larger or smaller value by chance
  • Choose the significance value alpha – a typical value of 5
  • Decision rule (threshold) u determines false positive rate alpha
  • Choose u to give acceptable alpha under H0
  • Compute test statistics and compare it with voxel threshold or below
  • Voxel threshold – significant effect
  • When you get an effect assign P – which is the area of the curve corresponding to statistical u
25
Q

What is multiple tests?

A

• If we have 100,000 voxels, alpha=0.05  5,000 false positive voxels
- Scale the number of voxels
- No control over the number of false positives
• This is clearly undesirable; to correct for this we can define a null hypothesis for a collection of tests

26
Q

What is family-wise null hyppthesis?

A

Activation is zero everywhere

27
Q

What happens if a voxel null hypothesis at any voxel is rejected?

A

We reject the family-wise null hypothesis

28
Q

What gives a family wise error (FWE)?

A

A FP anywhere in the image

29
Q

What is the family-wise error rate (FWER)?

A

Corrected p-value

30
Q

What is the Bonferroni correction?

A

an adjustment applied to p-values that is supposed to be applied, when two or more statistical analyses have been performed on the same sample of data
• Conducting multiple analysis on the same sample of data

31
Q

What is Type 1 error?

A

Erroneously rejecting the null hypothesis with a statistical analysis, when the null hypothesis is in fact true in the population

32
Q

How can you test the null hypothesis?

A

with an independent samples t-test

33
Q

When do you reject the null hypothesis of equal means?

A

If the t-test p-values is less than .05.

34
Q

Whatt is the familywise type 1 error rate known to be larger than?

A

per analysis error rate (i.e. alpha = 0.05)

35
Q

What is the Familywise Error rate?

A

The probability of committing at least one type 1 error amongst two or more statistical analyses on the sample of data

36
Q

What is the Random Field Theory?

A

Consider a statistic image as a discretisation of a continuous underlying random field

37
Q

What is Euler characteristic?

A
  • Topological invariant
  • A number that describes a topological space’s shape or structure regardless of the way it is bent
  • It is commonly denoted by X
  • Compute at any threshold
  • At high threshold, the Euler’s characteristic is counting the number of blobs
  • If the threshold is really high, the Euler’s characteristic is either 1 or 0
38
Q

What is Fixed effect analysis (FFX)?

A
  • In order to test whether the obtained results are valid at the population level, the statistical procedure needs to consider that subjects constitute a randomly drawn sample from a large population
  • Inferences drawn from obtained results are only valid for the included group of subjects since the group data is treated like a virtual single-case study
  • Subjects are thus random quantities and the statistical analysis must assess the variability of observed
  • Simple model
  • lots of degrees of freedom
39
Q

What is disadvantage of FFX?

A
  • Large amount of data

- Assumes common variance over subject at each voxel

40
Q

What can be compared with Fixed Effects Analysis (FFX)?

A

The group effect to the within-subject variability

It is not an inference about the population from the subjects were drawn

41
Q

What is compared with Random Effects Analysis (RFX)?

A

Compare the group effect to the between-subject variability. It is an inference about the population from which subjects were drawn

42
Q

Fixed-effects

A
  • Only one source of random variation (over session)
  • Measurement error
  • Within-subject variance
43
Q

Random Effects

A
  • Two sources of random variation
  • Measurement errors – within-subject variance
  • Response magnitude (over subjects) – Between-subject variance
44
Q

The summary statistics approach is exact if for each session/subject

A
  1. Within-subjects variances the same

2. First level design the same (e.g. number of trials )

45
Q

What is summary statistics approach?

A

Robust against typical violations