9 - The Man Who Set Back Machine Learning (Not Really) Flashcards
(46 cards)
What is deep learning?
The process of training neural networks that have three or more layers (one input layer, one output layer, and one or more hidden layers).
Who is George Cybenko?
A professor of engineering at Dartmouth College known for his work on neural networks.
What significant event occurred in 2017 related to deep learning?
A summer school on deep learning in Bilbao, Spain, attended by nearly thirteen hundred people.
What is the universal approximation theorem?
A theorem that shows a neural network with just one hidden layer, given enough neurons, can approximate any function.
What did Cybenko’s landmark paper demonstrate?
It proved that a neural network with one hidden layer can approximate any function.
What did Minsky and Papert’s book ‘Perceptrons’ conclude?
Single-layer neural networks have limitations and multi-layer networks are likely useless.
What is backpropagation?
A training algorithm for neural networks that allows for the training of multi-layer networks.
Who were the authors of the seminal 1986 backpropagation paper?
David Rumelhart, Geoffrey Hinton, and Ronald Williams.
What is the role of a single-layer perceptron?
It takes input values, calculates weighted sums, and produces an output based on a thresholding function.
What is the perceptron training algorithm used for?
To train a single-layer neural network by finding optimal weights and biases.
What differentiates a deep neural network from a single-layer network?
A deep neural network has multiple weight matrices due to the presence of hidden layers.
True or False: The perceptron training algorithm can be used for networks with hidden layers.
False.
What does training a neural network involve?
Finding optimal values for the weight matrices to approximate a desired function.
Fill in the blank: Cybenko’s work is often associated with _______.
delaying deep learning by twenty years.
What can a function represented by a neural network achieve?
It can represent a decision boundary or perform regression.
What is an activation function?
A function that determines the output of a neuron based on its input.
What is the sigmoid activation function?
A smooth function that transitions from almost 0 to almost 1.
What type of neuron did Cybenko use in his proof?
A nonlinear neuron based on the sigmoid activation function.
What is the significance of having multiple weight matrices in a network?
It allows the network to be classified as a deep neural network.
What does the term ‘AI winter’ refer to?
A period of reduced funding and interest in artificial intelligence research.
What is a hidden layer in a neural network?
A layer of neurons that is not directly exposed on the output side.
What is the output of a bipolar neuron?
+1 or -1.
What activation function is used in Cybenko’s neurons?
The sigmoid activation function, a(z) = σ(z)
The sigmoid function smoothly transitions from almost 0 to almost 1.
What does the equation z = wx + b represent in the context of a neuron?
The weighted sum of the inputs plus the bias term
Here, z is calculated using the weight vector w and bias b.