FINAL Flashcards

1
Q

2 higher order theories of consciousness (HOT)

A
  • higher-order perception
  • higher-order thought
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

higher-order perception

A

idea that consciousness arises when you acknowledge yourself perceiving other things

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

higher-order thought

A

idea that humans are aware of thought processes, and this brings about consciousness

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

3 criticisms for higher order theories

A
  • we are aware of stimuli, not thoughts about stimuli (meta-thought is not necessary for consciousness)
  • doesn’t take conscious action into account
  • no neuro basis
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

global workspace theory (GWT)

A
  • input processors compete for attention (similarly to pandemonium -> loudest will enter consciousness/workspace)
  • (conscious) global workspace broadcasts input to other brain areas for voluntary action
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what 3 things does GWT account for

A
  • intentional action
  • information retention
  • planning and creative thought
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

global neural workspace theory (GNW)

A

used fMRI to map brain areas with different functions, these propagate signals across to motor systems to organize voluntary action

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what structure is thought to account for consciousness and why according to GNW

A

pyramidal neurons because of widespread structure that can connect many areas including PFC and temporal lobes for voluntary action

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

monism

A

body and mind made of one substance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

dualism

A

body and mind made of two distinct substances (possibly connect through pineal gland)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

functionalism

A

mental states are constituted only by their functional roles (i.e. accounts for multiple realizability); brain = hardware, mind = software

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

reductionism

A

break everything down into parts, no longer discuss the whole

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

advantage of reductionism

A

easy to test each individual part

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

disadvantage of reductionism

A

some things can’t be broken down

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

emergence

A

sum > parts e.g. janitor’s dream

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

rat memory experiment setup

A

play a tone, then shock the rat. if they remember, they should be afraid every time they hear the tone.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

how to prevent a rat from learning new things

A

administer drug while / near the time of the initial shock

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

how to prevent a rat from remembering things

A

administer drug during recall, it will alter the original memory

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

can the rat memory experiment be done in humans and is it effective

A

yes, drug was used with PTSD patients and was somewhat effective, it makes the painful memory less painful

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

best way to preserve a memory in its original form

A

don’t recall it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

why can we never be 100% certain about material reality (& order of processing pipeline?)

A

real world -> senses -> processing -> awareness; reality is always filtered through senses and processing therefore never 100% objective

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

cartesian theatre

A

example of higher-order perception (perceive yourself perceive something)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

problem with cartesian theatre

A

if homunculus is another person inside your head, what about their consciousness?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Dennett multiple drafts model

A

different senses have different processing streams where they process things intermittently, and objects can reach ‘fame’

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
'fame' (multiple drafts model) & what is it related to
related to GWT; something that reaches awareness when not being processed, and is broadcast to other brain modules
26
'draft' (multiple drafts model)
there can be multiple version of one stimulus in one stream due to lots of processing
27
can stimuli continue to be processed after they reach fame
yes
28
David Eagleman task summary
- monitor motor cortex - high specificity watch - any time you feel urge to press button, note the time - up to 2 seconds before reporting feeling urge, brain has already made the decision for you (without your awareness) therefore do we have free will?
29
eagleman democracy of mind
rivalling streams competing for attention (similar to pandemonium
30
3 features of eagleman democracy of mind
redundancy, competition, emotional + rational processing
31
redundancy of democracy of mind
each stream is processed by multiple competing systems
32
competition of democracy of mind
e.g. pandemonium, which stream is loudest
33
emotional + rational processing of democracy of mind
both kneejerk processing and more rational
34
what does IIT slicing allow for
seeing which systems are affected lets you determine which systems are dependent on each other vs independent
35
what is phi value when slicing
number of dependent subsystems
36
what does a high phi value mean (theoretically)
more consciousness
37
negatives of IIT (5)
- panpsychism - hard to calculate - possible to 'hack' - just an opinion - no definition for consciousness but now we have a number?
38
structural brain imaging (2) plus 2 examples
- shows anatomy - used for tumors, strokes, lesions. etc - CT and MRI
39
functional brain imaging (2) plus 3 examples
- shows blood flow / electricity - used during experiments and diagnosis - fMRI, PET, EEG
40
Computed Axial Tomography (CAT / CT) scan
slice by slice through brain (white = bone)
41
MRI
machine shoots radio waves at tissue, only H responds, creates excitation, H stores info, then measure H activity (white=bone)
42
CT advantages over MRI (3)
- better spatial res - cheaper - faster
43
MRI advantage over CT
- better contrast
44
fMRI
(multiple scans) oxygenated blood -> processing -> deoxygenated blood areas where blood is being deoxygenated light up (blood oxygen level-dependent (BOLD) signal)
45
Positron emission tomography (PET) scan
(rainbow scans)inject radioactive glucose, brain activity uses sugar -> see glow more
46
fMRI advantages over PET (3)
better spatial and temporal res, no radioactivity
47
PET advantages over fMRI (3)
- faster - quieter - cheaper
48
EEG advantages (5)
- fast - cheap - safe - direct relation to brain activity - good temporal res
49
what type of scan do we not have yet
full CNS scan
50
how do diff types of memory differ (3)
capacity, duration, content
51
2 types of sensory memory
iconic (visual = 250-200ms) and echoic (auditory = several secs)
52
why no olfactory, gustatory and tactile sensory memory
difficult experimental protocol
53
normal capacity of sensory memory
4-5 items e.g. letters
54
what did Sperling find sensory memory capacity could be extended to with training
9-12 items e.g. letters
55
what happens to sensory memory with a 1s distractor (masking)
removes majority of it (back down to like 3 items)
56
what happens to sensory memory with a cue to indicate direction (too short for conscious awareness)
restores performance
57
how do blinking vs blank screen affect sensory memory
blinking = disrupts performance blank screen does not
58
working memory duration without rehearsing
18s
59
working memory capacity (& how to improve it)
7 +- 2 chunking
60
3 types of coding for working memory
- acoustic (e.g. get confused because stimuli sound the same) - semantic (e.g. get confused because categories have similar meanings) - visual (e.g. rotation tasks are processed degree by degree)
61
is working memory scanning done in serial or parallel, and how do we know (in lab) (2) & caveat to these results
- RT linearly correlated to set size (7+-2) - we don't terminate as soon as we find the number = exhaustive (not serial) * perhaps only parallel processing because we want to do well in lab
62
2 types of LTM
explicit/declarative and implicit
63
2 types of explicit/declarative memory
semantic and episodic
64
implicit memory =
procedural
65
why is assessing LTM capacity hard
must max it out = how? - hard also because of memory reorganization
66
LTM duration (3 phases)
1. rapid decline over first 3y without reinforcement 2. stable at 75% for ~25-30 years 3. another decline, possibly due to general cognitive decline
67
how does better learning/memorization affect LTM duration
higher starting point, but curve stays the same
68
coding explicit/declarative memory
various locations across cortex (distributed representation)
69
coding implicit memory
production (e.g. if then rules) in cerebellum
70
what did Lashley find about LTM location (engram)
takes longer to retrain a rat depending on the size of the brain chunk that was removed, but rat does not forget
71
equipotentiality (Lashley)
brain areas can take over for each other after damage
72
Hebb rule
neurons that fire together wire together
73
long-term potentiation (LTP)
high freq = more receptors develop on receiving neuron
74
HM (procedure and result)
removal of hippocampus = anterograde amnesia (no new memories), but all memories up to that point are intact
75
hippocampus function
memory consolidation
76
why is cognitive science reverse-engineering
we are trying to figure out how an already existing thing works
77
machine definition
any cause-effect system
78
4 features of computation (and what is it ultimately)
- rule-based - shape-based - implementation-independent (i.e. multiply realizable) - semantically interpretable aka symbol manipulation
79
weak Church/Turing hypothesis
turing machine can compute anything that can be computed by a general-purpose digital computer
80
strong Church/Turing hypothesis
turing machine can compute anything that can be computed
81
what is mathematics fundamentally
syntax/semantics (manipulating shapes based on rules)
82
computationalism (strong AI)
'cognition is computation' (not really true)
83
what premise must be adopted because of computationalism's definition
that computation is implementation-independent
84
Searle's Chinese Room Argument (cognition is not computation)
even if you memorize and execute the whole rule book, you do not understand chinese
85
Searle's periscope (the implementation-independence of computation)
no implementation of the rule book leads to understanding mandarin
86
turing test hierarchy
t1 = toy (regurgitating patterns) t2 = verbally indistinguishable e.g. chatgpt t3 = verbal + robotic t4 = verbal + robotic + neuro
87
which level of turing hierarchy is disputed by Searle's Chinese room argument
t2
88
what level does Harnad think is correct
t3
89
symbol grounding problem
can look up mandarin symbols indefinitely but if you never have a referential meaning for the symbols the cycle will go on indefinitely
90
minimal grounding set
the 1000 words that can be used to define every other other word in the dictionary
91
how does minimal grounding set get its meanings (3)
- direct sensorimotor grounding (DSG) (e.g. trial and error in real world) - indirect verbal grounding (IVG) (e.g. describing the sensorimotor stuff)
92
what is a powerful way of grounding new words, but what is the issue with it
language; only works if you understand the words being used
93
why is computationalism the cogsci dominant theory
because it allows for everything to expressed as an algorithm
94
how are real neurons similar to fake neurons
lots of inputs converge on a neuron/'black box' to give 1 output
95
what is first step after input in artificial neuron
weigh each input
96
what comes after weighing each input in artificial neuron
sum of all inputs x weights
97
what comes after the sum of inputs x weights in artificial neuron
bias = add 1 number to every value to simulate base excitation level
98
what comes after bias in artificial neuron
activation function e.g. threshold
99
what is the implementation of an artificial neuron
dot product aka matrix multiplication (multiple input, outputs, weights, and biases)
100
what is input x weight in neuron implementation called
parameter
101
how mant parameters does chatgpt have
1.76 trillion
102
what is computer core like intel good for
good for complex math (smart but few)
103
what is each computer graphic core good for
very basic math (dumb but many)
104
what is sigmoid function an analogy for in artificial neurons
neuron threshold + maxing out
105
ReLU (rectified linear unit)
1 to 1 ratio of increase (above 0 on x-axis), output of 0 (below 0 on x-axis)
106
what do OR, AND, NOT functions have in common
each component just looks at its inputs, then transmits a 0 or 1 signal
107
perceptron model
same as neuronal model (inputs, weights -> output)
108
perceptron update rule (2)
small delta = desired output - actual output big delta = random small value * small delta * input * must calculate big delta separately for each weight in perceptron model
109
what does big delta mean
how much you should change each weight in the perceptron model to achieve desired output
110
why does applying big delta not always lead to exact desired output
because of the small random value (E)
111
what type of learning is applying big delta
supervised learning
112
multi-layer perceptron (MLP) / dense neural network
every layer is connected to every element in the next layer
113
'deep' neural network'
contains hidden layers that are not directly trained
114
NN training (general) (5)
- randomly initialize all weights - put input through model (feed-forward), receive predicted output - calculate loss (desired output - predicted output) - backpropagate loss through network - update all weights and biases
115
convolutional kernel
pattern it looks for; allows for feature search over a whole image and avoid localizability
116
how does convolution work
multiple neurons (e.g. 3x3) for 1 single pixel = convolved feature, repeat for whole image
117
AlexNet concept
same image can go through multiple convolutional kernels at the same time to look for more than 1 feature
118
physical symbol system
a set of entities (symbols) that are physical patterns that can occur as components of another type of entity called an expression/symbol structure
119
what are all rules in the brain theoretically saved as
expressions
120
how are concepts stored in the brain
combinations of many features
121
simon's symbolic model of emotion
(in higher level system) start -> physical symbol system (expression) -> done
122
what does simon's symbolic model of emotion not account for (4)
- classification of emotions - which emotions are more important - physiological markers - neuronal basis of emotion
123
how can simon's symbolic model. of emotion reprioritize based on emotions / urgency
CNS can cause interruption
124
how did Ekman find basic emotions
tested Papua New Guinea people of fore culture on emotion identification
125
ekman's 6 basic emotions
fear anger surprise sadness happiness disgust
126
2 claims about ekman's emotions
- distinct emotions with distinct physiological features - evolutionary functions (hardwired)
127
2 criticisms on ekman's basic emotions
- link between emotion and physiological response? - sociological impact necessary for proper development?
128
how do emotions differ (2)
duration and role
129
axes of russel's circumplex model of emotions
valence (goodness) and arousal (engaging) (2D scale, arranged in a circle)
130
Adolph & Anderson modern alternative to emotions
7 dimensions to create interspecies framework of emotions
131
7 dimensions of Adolph & Anderson emotion framework
- scalability: varying intensity - valence: pleasantness - persistence: outlast stimulus - generalization: specificity to stimulus - global coordination: engage whole organism - automaticity: how challenging to control - social coordination: social functions
132
appraisal theory of emotions
emotions lead to change in the perception of the environment (how emotions relate to cognition)
133
how an emotional episode is created (3)
- triggering stimuli and context interact to form perception - perception creates somatic and neural responses as well as cognitive evaluation/appraisal and emotional feelings - leads to behavioural / verbal responses
134
5 tools for studying emotions
- neural responses - somatic responses - affective responses - genetic tools - lesion studies (temp and permanent)
135
3 examples of genetic tools to study emotions
knockout experiments, optogenetics, pharmacogenetics
136
can we use language without communicating
not really
137
Paul Watzlawick's Axioms of communication (5)
- one cannot not communicate - communication is diff between diff people - communication is punctuated - communication involves digital + analogic modalities (verbal + non verbal) - communication can be symmetrical or complementary
138
Clark's language characteristics (5)
- communication - arbitrary (e.g. why is 'truck' a truck) - structured (syntax rules) - generative (infinite) - dynamic (constantly adding new words)
139
3 types of linguistic representation
auditory, visual, haptic (braille)
140
language processing pipeline (4)
phonemes, morphemes, syntax and semantics, pragmatics
141
phonology (3)
sounds of letters, IPA, spectrogram
142
coarticulation
phonemes modify each other
143
144
why is absence of freq on spectrogram not good for identifying word boundaries
occur within words too
145
temporal induction
strong top-down influence on phoneme perception
146
2 types of morphemes
stem and bound morphemes
147
how does auditory word recognition occur
all possible words get activated until stimulus narrows down options
148
how does written word recognition occur
we don't read letter by letter, we fixate and process the rest in our periphery
149
what causes increased word processing time (2)
- diff phonemes e.g. pint vs mint - double meanings
150
garden-path sentence
immediate meaning sounds wrong, must reparse and get to different outcome (include syntactic ambiguity)
151
Chomsky vs Lakoff
chomsky = syntax lakoff = semantics
152
how are all sentences represented according to lakoff
pictograms
153
5 pragmatic features
-assertives -directives -commissives (commit to later action) -expressives (about mental state) -declaratives
154
what do decision trees involve
lots of arbitrary biases
155
2 features of expert systems
explainable, can be hand-crafted
156
2 disadvantages of expert systems
difficult to handcraft, falls apart with large data/edge cases/nonlinear correlations
157
what is visual hierarchical processing similar to
CNN processing
158
3 CNN features
- sparse connectivity (every input does not connect to every output) - shared weights - invariance under translation (look for 1 feature across whole image)
159
autoencoders
encodes itself to learn compressed representation; dense encoded data reconstructed through backpropagation
160
what do smaller bottlenecks lead to in CNN autoencoders
worse reconstruction
161
latent space interpolation
reconstructing interpolated vectors allows you to see what comes between the 2 start and end vectors
162
latent space arithmetic
allows you to generate data that was not part of data set e.g. smiling woman - neutral woman + neutral man = smiling man
163
why must we do arithmetic on latent space rather than image itself
image - image could lead to 2 noses, all background space, etc.
164
why are some word embeddings e.g. dictionary position of a word not useful
because not meaningful
165
why is latent space arithmetic with words sometimes problematic
can lead to biases
166
recurring neural nets (RNN)
loop model into itself
167
what does RNN involve
sliding window (3 words as input -> following word as output)
168
what do RNNs allow for (in theory)
learn truth value of certain statements
169
what makes RNNs better
storage
170
why does storage help RNN
because otherwise separator characters have no data to draw a conclusion from
171
neural machine translation with encoder and decoder RNN
encode until 'end token', then decode and output
172
2 RNN types
GRU cells and LSTM
173
GRU cells
gated recurrent unit, only short term memory
174
LSTM
long short term memory; built in mechanism so you cannot delete stuff from beginning
175
problem with predictive text
often causes cyclic phrases if only pick #1 option
176
how to solve predictive text problem
add randomness i.e. pick an option in top 10
177
does looking at character frequency in isolation lead to language models
no (e.g. sd n oeiam etc.)
178
bigrams
co-occurences of letters/words (e.g. on inguman ise forenoft etc - not real words but resemble it more)
179
how to get good predictive text
use word bigrams/trigrams/n-grams to form word sequences
180
cosine similarity
cosine of angle between 2 vectors = quantifiable similarity between 2 things (dot product between 2 vectors)
181
person embedding
turn big 5 scores into vectors, fine cosine similarity, if vectors point in roughly same direction, people should get along better
182
what can be done with vector values (similar to latent space arithmetic)
make a slider out of each value in a vector, see what meaning each specific number corresponds to e.g. square 4 = personhood
183
proxy tasks
know how many of each object there is, ask something like "what is there the highest quantity of" - teaches system to count and rank things implicitly
184
CBOW (continuous bag of words)
(proxy task to learn word embeddings) - ask system to fill in blank (implicitly teaches system similarities between concepts e.g. king and queen can both sit on throne)
185
what type of learning are proxy tasks
supervised
186
GPT context window
number of words a transformer looks at before giving output; the more words you look at, the more contextual questions you can answer
187
what does chatgpt use instead of word embeddings
tokens (allows for certain characters to be grouped together therefore system can run better)
188
what must transformers learn
correspondence
189
key-value storage
key = concept value = explanation query = question about a concept (answer should be value)
190
how to execute key-value queries (6)
1. calculate word embedding (e.g. make a vector for word 'is') 2. turn vector into 3 vectors: query, key, value 3. make key and value vectors for 'sky' and 'the' 4. run query of 'is' against its own key 5. also run the query of 'is' against keys of 'the' and 'sky' (cosine similarity test) 6. highest cosine similarity will be the value (in this case, probably sky; value of sky=blue therefore output blue)
191
narrow/weak AI
good at one task (possibly better than humans)
192
artificial general intelligence
good at all tasks, can learn new tasks, can learn anything a human can and do it better
193
why do datasets matter for transformers
can learn different things from different datasets
194
problems w datasets
biases
195
reinforcement learning from human feedback
add training data into the model (feedback on how good response was)
196
problem w reinforcement learning from human feedback
sparse reward