Lecture 11 Flashcards

(36 cards)

1
Q

machine learning

A

algorithms that modify themselves through experience and exposure with data. These are statistical equations that adjust their parameters to better classify and predict data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

artificial intelligence (AI)

A

algorithms applied to solve novel problems. They utilize algorithm-
refinement from machine learning to understand patterns of data, and then applies a solution in a
novel setting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

domain specific hypothesis

A

there are specific brain regions involved for processing specific features

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

spreading activation hypothesis

A

Activation spreads across interconnected brain regions depending on the features being processed.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

supervised learning

A

involves training algorithms by providing datasets with predefined labels. The algorithms attempt to find the best ways to differentiate between these classes of labels.

For example:
what weights can be assigned to highlight the main differences in patterns of activity for natural versus
human made objects?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

unsupervised learning

A

involves providing data without class labels. The ML algorithm attempts to
discover the maximum separability between the patterns in the data.

For example: what weights can be
assigned to cluster this data into two different classes with the most separation?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

testing data

A

a novel set of data without class labels presented to the ML algorithm
after training. The ML algorithm attempts to “predict” a class label for the data based on the learned patterns of activation from the testing data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

types of machine learning

A

(1) support vector machines (SVM)
(2) k-nearest neighbour (kNN)
(3) neural networks (NNs)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

training data

A

consists of the data with class labels (or without, if unsupervised). The
ML algorithm devises the optimal weights to differentiate the classes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

differences between regression and machine learning

A
  1. Machine learning can take advantage of giant swaths of data, taking in large numbers of
    inputs, finding patterns, and providing predictions that maximize accuracy… but so can
    hierarchical regression models
  2. Machine learning can optimize its outputs as it gets more experience with data… but so can
    higher level Bayesian statistics
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

support vector machines (SVM)

A

divides data in multidimensional space into its two classes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

k-nearest neighbour (kNN)

A

creates a decision space based on vectors to neighbouring points

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

neural networks (NNs)

A

a collection of artificial nodes or “neurons” that connect to one another and receive, process, and transmit data, and whose structure is
inspired by biological brains and thus aims to mirror their higher level processing skills

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

hyperplane

A

divisor in SVM
a multidimensional plane that aims to best divide the data into two class labels

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

support vector

A

the points that touch the hyperplane, and define the position of the hyperplanes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

margin

A

the space between the two hyperplanes that separate the data. The larger the margin, the more
accurate the model’s predictability

15
Q

maximum-margin

A

the largest possible distance between the hyperplanes separating the data

16
Q

how does k-nearest neighbour (kNN) work

A

the class identity of a given test point is determined by its closest distance
(i.e., its neighbours) to k number of other points. Training involves calculating vectors between values in
multidimensional space
k is typically a small number (e.g., 3). If k = 1, then the decision of a test point is simply its closest neigbour

16
Q

process of SVM

A
  1. Data is collected (e.g., brain imaging data when people look at
    human made or natural objects)
  2. Labelled training data is fed into the SVM algorithm (X × Y x Z x T is
    organized into a single array, labelled as human made or natural)
  3. SVM fits a hyperplane to the data through its training. The hyperplane divides the data into two class labels
  4. A new set of testing data, unlabeled, is fed into the SVM model
    which attempts to predict the class labels for the test data. Accuracy
    is determined as the percent that was correctly labelled
17
Q

decision space

A

can be generated by integrating the closest k points in multidimensional space

18
Q

k-fold cross validation

A

where you combine all iterations with a given subsample of data (k) is left out for testing

19
Q

NNN layers

A

input layer receives data and begins initial processing.
The output layer outputs the final set of values or set of values

20
Q

simple NNN

A

one with a maximum of three layers. The input layer is connected directly to the output layer and may have one hidden layer (where additional computations takes place)

21
Q

deep NNN

A

ave more than one hidden layer. These are often considered black boxes

22
Generative adversarial networks (GANs)
involve two NNs, one that generates an image (the generator), and another that compares it to real images to determine whether it can distinguish between the real and generated one (the discriminator)
23
classifications of neural networks
1. Recognition/classification to learn and identify class labels of information, images, data, etc. This can involve recognizing images such as scenes, people, or objects, or identify things like cancer on a CT scan 2. Generation to produce novel data or images. This is a feature that other ML algorithms like SVM/kNN do not have
24
narrow artificial intelligence (NAI)
generative ai that is trained on data for specific outcomes and can only provide predictions based on its internal boundaries
25
Artificial generalized intelligence (AGI)
can train from a wide variety of contexts and apply its understanding in novel and unseen situations. It can learn from one setting and generalize to a completely different setting. This is on par with the human ability to generalize
26
tokenized
words are tokenized; where they are represented as coordinates in a multidimensional space. A given sentence is thus represented by its relationship between these points in space. ChatGPT then tries to predict the appropriate response based on the structure of these tokens
27
chatgpt
Generative Pre-trained Transformer. These are large language models trained in unsupervised data to generate text. The transformer is what tokenizes words, and these vectors are understood in the context to other tokenized words
28
issues with machine learning
1. Is it picking up actual categorical differences between classes, or some lower-level attribute? 2. Some machine learning (especially DNNs) are black boxes 3. Can normal NHST explain the results? 4. Slight discrepancies in the data can bias the algorithm 5. Overfitting
29
5. Overfitting
A model tries to fit the input data so well that it essentially loses its generalizability to other data
30
4. Slight discrepancies in the data can bias the algorithm
The very fact that the human-made mean is slightly larger, or have a slightly different distribution, may be enough to bias a machine learning algorithm. Often, we have to do lots of transformations to standardize and normalize datasets being compared
31
1. Is it picking up actual categorical differences between classes, or some lower-level attribute?
Differences in relative pixel size, lighting, colour, contrast, illumination, shape, texture, perceived size, perceived weight, emotional valence, colour pallet, familiarity, etc, can all contribute to “differences” between classes. Researchers implementing ML algorithms need to often work harder on standardizing differences between classes than they do on the actual machine learning component. This can greatly restrict and impact the validity of these stimuli
32
2. Some machine learning (especially DNNs) are black boxes
Yes, we can accurately predict information, or even create DNNs that mimic human functions like vision or speech, but do not provide any additional information about what is going on in the brain or with cognition. DNNs can get extremely complex with recurrent networks. It becomes impossible to visualize information at any given layer
33
3. Can normal NHST explain the results?
Fancy applications of fancy algorithms don’t add any explanatory value and can often introduce more problems than solutions. T