3. Visual Healthcare Analytics Flashcards

1
Q

what is parallel coordinate

A
  • visual of multidim data
  • given NxM table with N patients and M clinical var, a line chart is generated by displaying M equally spaced vertical axes with individual ranges
  • can then filter and explore correlations between vars for patients
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

draw a parallel coordinate graph

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what is a chord visualisation plot

A

a diagram illustrating the connection between different variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

draw a chord visualisation plot

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what are 2 dim reduction algos

A
  • pca
  • t-SNE
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is t-SNE

A

an algorithm that calculates similarity via the high dim and low dim space. it computes the distance between instances in both spaces and tries to optimise these similarity measures using a cost function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what is pca

A
  • unsupervised linear dimensionality reduction
  • visualisation technique
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

how is pca calculated at a high level (variance)

A

the greatest variance by some scalar projection becomes the first PC (coordinate), and the second greatest becomes the second PC etc etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

how can we determine the amount of preserved information after pca is applied

A

variance - compare that of all original dimensions, with the variance of the reduced dimensions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

after performing pca, two variables have a variance of 1.46 (A), and 0.2 (B). the entire dataset itself has a variance of 2.06, what can be inferred by this

A

we can infer that variable A alone can explain most of the information of the output predicted by the two variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

how is pca calculated

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

how are the number of pca dimensions measured

A
  • consider using a scree plot. it shows the variance explained by each PC, based on the number of PCs used
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

draw a scree plot

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

when should we not use PCA

A

pca is linear because when calculated, it is projected as a linear vector [finish this]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what to use instead of pca

A

if it’s not a linear transformation, consider distributed stochastic neighbourhood embedding (t-SNE)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

how is t-sne calculated

A
  1. for every point in 2D space, centre a Gaussian distribution over that point (this denotes the similarity by probability, of the two points)
  2. repeat step 1, this time with a Cauchy/student t-distbn (these are the set of prob. for the low dim space)
  3. now map the high dim and low dim using a loss function to minimise the distance using Kullback-Liebler divergence (KL)
  4. finally, use gradient descent to minimise the KL cost function
17
Q

what is the advantage of the student t over Gaussian

A

student-t has heavier tails than normal, so it can account for less similar points (further distances)

18
Q

compare t-SNE & PCA

A
  • linearity
  • t-SNE doesn’t keep the original data structure while PCA does
  • t-SNE produces well separated clusters
  • t is computationally expensive for large datasets
  • PCA is consistent, t-SNE is stochastic and can produce varying embeddings
  • PCA requires no hyperparam optimisation
19
Q
A