Machine Learning Flashcards
(199 cards)
adversarial training
AT: “training a model in a worst-case scenario, with inputs chosen by an adversary”
Adversarial training is often used to enforce constraints on random variables
GAN
GAN is a generative model that learns the probability distribution (or data distribution) of the training examples it is given.
From this distribution, we can then create sample outputs. GANs have seen their largest progress with image training examples, but this idea of modeling data distributions is one that can be applied with other forms of input
=> the key mathematical tool GANs give you is the ability to “estimate a ratio”
==> GANs are generative models that use supervised learning to approximate an intractable cost function by estimating ratios
discriminator network
t
convolutional neural network
a convolutional neural network (CNN, or ConvNet) is a class of deep, feed-forward artificial neural networks that has successfully been applied to analyzing visual imagery
deconvolutional neural network
t
backpropagation
In principle, all backpropagation does is (stochastic) gradient descent -> This converges to a local minimum, which are often enough surprisingly good
Backpropagation is a method used in artificial neural networks to calculate a gradient that is needed in the calculation of the weights to be used in the network. It is commonly used to train deep neural networks, a term used to explain neural networks with more than one hidden layer
inpainting
In the digital world, inpainting (also known as image interpolation or video interpolation) refers to the application of sophisticated algorithms to replace lost or corrupted parts of the image data (mainly small regions or to remove small defects)
damage and repair strategy
t
pretext task
t
autoencoder
is an artificial neural network used for unsupervised learning of efficient codings (ie feature learning)
The aim of an autoencoder is to learn a representation (encoding) for a set of data, typically for the purpose of dimensionality reduction
low-level statistics
e.g. unusual local texture patterns
variability
t
epoch
An epoch is one complete presentation of the data set to be learned to a learning machine. Learning machines like feedforward neural nets that use iterative algorithms often need many epochs during their learning phase.
adversarial examples
inputs to machine learning models that an attacker has
intentionally designed to cause the model to make a mistake; they’re like optical illusions for machines
-> in computer vision: usually an image formed by making small perturbations to an example image from a dataset
tensor
multidimensional array
GAN
The dueling-neural-network approach has vastly improved learning from unlabeled data.
spatial resolution
Spatial resolution is a term that refers to the number of pixels utilized in construction of a digital image. Images having higher spatial resolution are composed with a greater number of pixels than those of lower spatial resolution.
saccade
is a quick, simultaneous movement of both eyes between two or more phases of fixation in the same direction
no free lunch theorem
“averaged over all possible data-generating distributions, every classification algorithm has the same error rate when classifying previously unobserved points”
in other words, the most sophisticated algorithm we can conceive of has the same average performance (over all possible tasks) as merely predicting that every point belongs to the same class.
Fortunately, these results hold only when we average over all possible datagenerating distributions.
But If we make assumptions about the kinds of probability distributions we encounter in real-world applications, then we can design learning algorithms that perform well on these distributions
==> the no free lunch theorem implies that we must design our machine learning algorithms to perform well on a specific task
hypothesis space
e.g. linear regression has a hypothesis space consisting of the set of linear functions of its input
regularization
we can regularize a model that learns a function f(x; θ) by adding a penalty called a regularizer to the cost function
Regularization is any modification we make to a
learning algorithm that is intended to reduce its generalization error but not its training error. Regularization is one of the central concerns of the field of machine learning, rivaled in its importance only by optimization
loss function
when we minimize the objective function, we may also call it the cost function or loss function
objective function
the function we want to minimize or maximize, also called criterion
=> when we are minimizing the objective function, we may also call it the cost function or loss function
squared L^2 norm
can be calculated simply as x^Tx
-> is more conenient to work with mathematically and computationally than the L^2 norm itself