Hearing Flashcards
(41 cards)
What is sound
Pressure waves in the air
It is a sequence of pressure waves which propagate through this elastic net.
What are the properties of air
Air is ‘elastic;’ if you try to compress it, it will push back.
Air has both elasticity and mass; so we can imagine air as being made up of little ‘lumps of air’ where each lump is connected to the next lump by an elastic spring.
How is the spectrum of the sound determined
Any complex sound wave can be produced by adding together sinusoidal sound waves (pure tones)
The amplitudes and frequencies of these components determine the spectrum of the sound.
What is the head-related transfer function (HRTF)
When sound is transformed from the external ears (pinnae) from free-field sound to the sound at the ear drum.
HRTF adds localisation information into the signal (this is why ears are shaped as they are).
What happens in the middle ear
It contains a set of intricate, interconnected bones: stapes, incus and malleus.
These amplify the motion of the ear drum into pressure waves transmitted via the oval window into the fluid-filled cochlea (in the inner ear)
What happens in the inner ear (cochlea)
Basilar membrane inside the cochlea vibrates according to the frequency structure of the sound.
Hair cells attached to the basilar membrane then transduce this mechanical signal into action potentials in auditory nerve.
What is the tonotopic map in the cochlea
It is a map of sound frequency that is created by the basilar membrane through its vibrations.
This is achieved through its structure: wide and stiff at the base, narrow & loose at apex.
The apex vibrates for a low frequency while the brave vibrates for a high frequency.
The cochlea separates out the different frequency components of a complex sound.
Thus, different auditory nerve fibres represent different frequencies = place code of frequency
What is phase locking
When auditory nerve fibres produce spikes that are phase-locked (synchronised) to the vibration of the basilar membrane.
What are the two ways that frequency components of sound are represented
Place of activity - tonotopic mapping or place coding
Timing of activity - temporal coding in which sounds with higher frequencies produce higher rates of synchronised firing.
What is the main task of subcortical processing
Sound source localisation utilising cross-ear differences in:
Sound wave amplitude
Sound wave arrival time
What are the properties of the visual nerve
The retina is 2-dimensional
Provides shape information
Provides colour information
Task for visual system to separate out visual objects relatively easy
What are the properties of the auditory nerve
Basilar membrane provides inofrmation on one dimension only:
Frequency content of the incoming sound wave
This sound wave is the the sum of all signals from around
Task for auditory system to separate out auditory objects is huge.
Subcortical pathway performs (at least) source localisation as part of the solution.
How is the auditory cortex organised
Core (C): primary fields (HG) receiving input from thalamus
Belt (B): secondary fields surrounding core
Parabelt (PB): secondary fields next to belt (STG)
Hierarchal organisation: activity progresses from C to B to PB
Parallel organisation: activity propagates along multiple C-B-PB paths
What happens in the cortical visual areas
Processing is distributed and specialised: each area represents specific type of information; orientation, colour, motion, etc.
What and where processing is achieved through two processing streams.
What do the auditory cortical fields do
Core fields respond to pure tones and complex sounds.
Belt and parabelt fields respond to complex sounds and noise, and less vigorously to pure tones.
No tonotopic organisation in parabelt.
However, there is little evidence of the kind of specialisation seen in visual cortex.
Which features do cells in the auditory cortex show selectivity for
Intensity
Bandwidth
Sound source location
What are the ‘what’ and ‘where’ processing streams
Anterior ‘what’ stream: specialisation in sound source identity (e.g. semantic meaning)
Posterior ‘where’ stream: specialisation in sound source location.
But this is a debated issue; what and where info is mixed
What are the properties of spectrotemporal receptive fields
Sharp and wide tuning to frequency
Tuning to sounds with complicated special structure
Tuning to sounds whose spectral structure changes
How is the auditory cortex modulated
By top-down effects which can be seen in discrimination tasks:
- Attention modulates STRFs (Fritz et al., 2003)
- Working memory shows up as sustained activity in
auditory cortex (Huang et al., 2016)
- • Perceptual learning changes mappings in auditory cortex
(Polley et al., 2006)
Describe perceptual learning in the auditory cortex
Polley et al (2006) - Rats trained to discriminate stimuli either according to pitch or loudness
Pitch discrimination training leads to expansion of frequency representation
Loudness discrimination training leads to expanded loudness representations
Thus, representations are modulated by top-down effects of learning
Describe surface recordings from cortex
Mesgarani & Chang (2012, Nature) presented sentences to human epileptic patients.
Surface recordings from auditory cortex
Method: stimulus spectrogram reconstruction from neural population responses + using machine learning classifier to decode stimulus from neural
response.
In the mixed condition, auditory cortex is responding as if the sentence being attended to is being presented alone.
What are some of the subcomponents of speech processing.
- Basic auditory processing of complex sound
- Distinguishing speech from non-speech
- Identifying speech sounds (e.g. vowels)
- Semantics at word, sentence, and discourse level
- Recognising prosodic aspects of speech
- Recognising individual voices, accents, & style of speech
- Recognising emotions
- Merging auditory and visual information
What is the traditional model of speech processing
Speech production occurs in Broca’s area in inferior frontal gyrus (IFG)
Speech comprehension occurs in Wernicke’s area in
posterior superior
temporal gyrus (pSTG)
But this no longer holds
Speech intelligibility
Scott et al. (2000) did a PET
study:
They contrasted responses to intelligible vs. non -intelligible speech stimuli
Larger responses to intelligible speech in left anterior STG
(so not Wernicke’s area)