Lecture 11 Flashcards
(36 cards)
machine learning
algorithms that modify themselves through experience and exposure with data. These are statistical equations that adjust their parameters to better classify and predict data
artificial intelligence (AI)
algorithms applied to solve novel problems. They utilize algorithm-
refinement from machine learning to understand patterns of data, and then applies a solution in a
novel setting
domain specific hypothesis
there are specific brain regions involved for processing specific features
spreading activation hypothesis
Activation spreads across interconnected brain regions depending on the features being processed.
supervised learning
involves training algorithms by providing datasets with predefined labels. The algorithms attempt to find the best ways to differentiate between these classes of labels.
For example:
what weights can be assigned to highlight the main differences in patterns of activity for natural versus
human made objects?
unsupervised learning
involves providing data without class labels. The ML algorithm attempts to
discover the maximum separability between the patterns in the data.
For example: what weights can be
assigned to cluster this data into two different classes with the most separation?
testing data
a novel set of data without class labels presented to the ML algorithm
after training. The ML algorithm attempts to “predict” a class label for the data based on the learned patterns of activation from the testing data
types of machine learning
(1) support vector machines (SVM)
(2) k-nearest neighbour (kNN)
(3) neural networks (NNs)
training data
consists of the data with class labels (or without, if unsupervised). The
ML algorithm devises the optimal weights to differentiate the classes
differences between regression and machine learning
- Machine learning can take advantage of giant swaths of data, taking in large numbers of
inputs, finding patterns, and providing predictions that maximize accuracy… but so can
hierarchical regression models - Machine learning can optimize its outputs as it gets more experience with data… but so can
higher level Bayesian statistics
support vector machines (SVM)
divides data in multidimensional space into its two classes
k-nearest neighbour (kNN)
creates a decision space based on vectors to neighbouring points
neural networks (NNs)
a collection of artificial nodes or “neurons” that connect to one another and receive, process, and transmit data, and whose structure is
inspired by biological brains and thus aims to mirror their higher level processing skills
hyperplane
divisor in SVM
a multidimensional plane that aims to best divide the data into two class labels
support vector
the points that touch the hyperplane, and define the position of the hyperplanes
margin
the space between the two hyperplanes that separate the data. The larger the margin, the more
accurate the model’s predictability
maximum-margin
the largest possible distance between the hyperplanes separating the data
how does k-nearest neighbour (kNN) work
the class identity of a given test point is determined by its closest distance
(i.e., its neighbours) to k number of other points. Training involves calculating vectors between values in
multidimensional space
k is typically a small number (e.g., 3). If k = 1, then the decision of a test point is simply its closest neigbour
process of SVM
- Data is collected (e.g., brain imaging data when people look at
human made or natural objects) - Labelled training data is fed into the SVM algorithm (X × Y x Z x T is
organized into a single array, labelled as human made or natural) - SVM fits a hyperplane to the data through its training. The hyperplane divides the data into two class labels
- A new set of testing data, unlabeled, is fed into the SVM model
which attempts to predict the class labels for the test data. Accuracy
is determined as the percent that was correctly labelled
decision space
can be generated by integrating the closest k points in multidimensional space
k-fold cross validation
where you combine all iterations with a given subsample of data (k) is left out for testing
NNN layers
input layer receives data and begins initial processing.
The output layer outputs the final set of values or set of values
simple NNN
one with a maximum of three layers. The input layer is connected directly to the output layer and may have one hidden layer (where additional computations takes place)
deep NNN
ave more than one hidden layer. These are often considered black boxes