What does the classical neuronscience approach allow us to do?
It allows us to CLASSIFY neurons according to:
1. Morphology → how to dendritic trees look like?
2. Function → more different functions deeper in visual pathway
3. Firing pattern → bursts, frequency
Etc.
*Every knob is a feature (many many knobs)
What is one of the importance uses of functional classifications of neurons?
(Classic approach)
Entorhinal cortex → Grid cells vs Border cells
Hippocampus:
- Place cells (being in a single spot in the room)
- Object-vector cells (respond to single object in that roome)
- Splitter cells (respond to being in a specific location but only when about to turn left)
What are advantages and disadvantages of the classic approach?
Advantages:
- Describes what individual neurons contribute to the computation
Disadvantages:
- Typically function is considered in specific setting → would have to consider ALL possible settings
- Small populations of neurons are considered → limited amount of cells you can record from
- circuits and mechanisms have to be deducted based of intuition from a very limited amount of cell information
What is the difference between deep learning and machine learning?
Machine learning:
Input → Feature extraction by a domain expert → Classification → Output
*Domain expert defines key features on top of which classification fits
Deep learning → Feature extraction + Classification → Output
*The network itself establishes the features and classifications directly from the data (in the learning process)
What are the “4 knobs” of deep learning?
(4 components on which the design is focused)
What are the different types of architecture a neural netowork can take?
1-2 are best for image inputs
1. Multilayer Perceptrons:
Every circle/neuron of 1 layer is connected to every neuron of both adjacent layers (not connected to other of the same layer)
- min 3 layers: Input → Hidden → Output
- Large number of weight parameters need to be trained
2. Convolutional Neural Network
1 nose detector goes through the whole picture (convolution), no need for different nose detectors for different areas (check for specific patterns over all the image → extract features)
What are the different cost functions of neural networks in deep learning?
*These are ways to learn/change parameters to improve the output
1. Unsupervised objective functions
- NO teacher
- For cross-modal consistency (read → write, Hear → talk) ~ generative consistency
- For future predictions (transformers trained to do that) → predict an image of a car moving in 2 secs, predict next work in a sentence
- may fails to discover properties of the world that are statistically weak but important for survival (need supervised for that)
*What cost functions does brain optimize?
*What do cost functions look like in the brain?
What are 3 ways of representing costs of neural networks in the brain?
What do we know about learning rules in the brain neural network?
*For the synapses that are potentiated
1. Changes in size of synaptic connections
2. Perforations caused by LTP
3. Multiple spines buttons (1 pre for multiple post)
To relate to neural networks:
Before training → all nodes have the same weight (all connected equally, parameters)
Training modulates the weight of different parameters/connections
Wtrained = WInitial + ∆W
What are types of datasets?
What are the main takeaways of the DL approach that are different from classical approach
Which future questions could the DL framework allow us to resolve?
Why is object recognition such a challenge for the brain?
*Difficult computational problem
1. Have to consider the infinit amount of ways an object can be shown to us and be able to distinguish it every time → different sizes, angles, colors, backgrounds, etc.
*Ventral visual pathway solves this problem
How long does it take for the brain to discriminate different objects?
Information gets to IT cortex ~ 100ms → this area discriminates different objects
~ 40ms → LGN
~ 50ms → V1
by 10ms every cortex higher
V1 → V2 → V4 → PIT → CIT → AIT
What are the first cells in the visual system to communicate via Action Potentials?
Retinal ganglion cells
~1.5 million in monkeys
Photoreceptors and Bipolar cells do not send APs
What are pinwheels in V1?
They are points around which if you turn (closely), you will encounter selectivity for all orientations
What different regions are found in IT?
What is population coding?
It is a way of looking at groups of neurons and their activity instead of single neurons
The ventral stream transforms this lined from being very curved → more linear as go higher up in cortical areas
*1 line is called the manifold of an object
For what type of information do you have to look at groups of neurons to learn about? (not found in single neuron’s activity)
3D scale, Z-axis rotation, height, width, perimeter
*response to these informations increase as we go up the ventral stream
→ They are category-orthogonal object properties (don’t fit in neuron categories)
How is memorability correlated
Memorability is dictated by the MAGNITUDE (spikes/sec) of the response of the specific set of neurons that always responds to the specific stimulus
→ not change in which neurons fire and the weight of each neurons in the global response
What was the name of the 1st neural net developped?
By who was it?
Multi-layer perceptron
Rosenblatt - 1958
*Uses the artificial neuron model
Who created the 1st artificial neuron model?
What was its principle?
McCulloch and Pitts (1943)
Dendrites → Dendritic tree connecting to soma/Synapse → Axon → Downstream synaptic terminal
Input → weighted parameters → threshold function → Output
The input to the threshold function is the sum of all the (weights of n)*(input given to dendrite n) → if it is over the threshold, then the neuron will fire AP
What is the structure of the multi-layer perceptron of Rosenblatt (1958)?
*First neural net
Retina → {Localized connections} → Protection area (A1) → Association area (A2) → Multiple Unit responses
*All other connections are random
- Each unit of each layer is connected to all the units of the previous/upstream layer
- Connections have different weights
If input 1 image of 10 pixels → each unit of A1 has 10 parameters → each unit of A2 has 100 parameters (*10)
What is the neocognitron?
It is a hierarchical, multilayered artificial neural network proposed by Kunihiko Fukushima in 1979
- Considers the fact that neurons of V1 are not connected to ALL neurons of LGN, same for V2, and V4…
- Considers that different units connect to different sets of units in the previous layer (ex: T-detectors)
- V1 neurons has smaller RF than V2 smaller than V4 (and more peripheral neurons have bigger RF than foveal)
- Layer have alternating S-cells (feature extraction) and C-cells (feature pooling for invariance).
*Precursor of CNNs