Perception 1: Low level vision Flashcards

(40 cards)

1
Q

What is the receptive field of a neuron?

A

Part in visual space where stimulus give rise to stimulating a neuron - receptive field of a neuron
Excite/inhibit ganglion cell

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How are ocular dominance columns (right or left eye dominant) arranged in relation to orientation columns? (columnar arrangement in V1)

A

Perpendicular

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a hypercolumn? (in V1)

A

A cortical processing module for a stimulus that falls within a particular retinal area

Formed from one set of ocular dominance columns and one set of orientation columns

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is feature detection theory of visual processing?

A

As you go through ventral stream, cells respond to larger and more complex stimuli
Hierarchy of cells

e.g.
V1 - edges and lines
V2 - shapes
V4 - objects
IT - faces

(receptive field size increases)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What two variables can affect spikes per second? (response rate in a particular cell)

A

Orientation and contrast

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the Fourier analysis framework?

A

Visual system deconstructs an image into discrete channels
Each channel conveys information contained in the image at a specific spatial scale and orientation

High spatial frequencies (low scale) for visual detail
Low spatial frequencies (high scale) for broad structure

Visual system then recombines them to form a coherent representation of the scene

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is coarse vs fine luminance?

A

Coarse luminance changes in an image - reveal large-scale structure

Fine luminance changes in an image - reveal small-scale detail

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What happens if you remove high vs low spatial frequency content from an image?

A

High removed = coarse (blurry)
Low removed = fine (detailed but no sense of bigger picture)

Info present at different scales within the same image

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does the mathematical theorem behind Fourier analysis state?

A

Any complex signal can be constructed from a set of simpler sinusoidal functions

So, vision can be broken down into simpler parts (frequency and contrast) to explain its complexity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does contrast sensitivity vary as a function of?

A

Spatial frequency

This is the contrast sensitivity function (CSF)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the contrast sensitivity function?

A

How well you can see detail across a full range of spatial frequencies

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

When is sensitivity (CSF) best and why?

A

Sensitivity is maximum between 2-5 cpd

Sensitivity is better at central range of spatial frequencies as you need less contrast in the image to resolve a pattern
At extremes of spatial frequency - more contrast is needed to resolve a pattern

More sensitive = you don’t need as much contrast to perceive something

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What did Blakemore and Campbell want to show with their spatial frequency adaptation experiment?

A

Show that someone’s sensitivity can change within a narrow range of spatial frequency as a result of adaptation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is contrast threshold?

A

The contrast of an image, below which the pattern looks homogenous (cannot see any detail)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What was Blakemore and Campbell’s (1969) spatial frequency adaptation experiment?

A

Participants set the contrast of an image (grating - stripes) so that they could just perceive luminance differences

Then, they introduced a high contrast adapting stimulus (clear stripes) for 60 secs
- This was at a specific spatial frequency
- Neurons responding to this dull and habituate, becoming less sensitive

Then, participants redo the first task (for contrasts at the same spatial frequency)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What happens to contrast threshold after adaptation? What does this show about neurons involved? (Blakemore and Campbell)

A

It increases, contrast has to be higher for people to perceive differences
Harder to see stimuli at a lower threshold

Higher threshold = easier to see stimuli

Shows that neurons for that spatial frequency habituated to adaptive stimulus and became less sensitive

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

After adaptation, how do threshold and sensitivity change in response to spatial frequencies similar to the adapting frequency?
What does this selective adaptation effect imply the existence of?

A

Threshold is increased
Sensitivity is decreased

This selective adaptation effect implies the existence of multiple, overlapping, spatial frequency channels

18
Q

What electrophysiological evidence supports conclusions from the spatial frequency adaptation effect - that there are multiple, overlapping, spatial frequency channels?

A

Contrast sensitivity functions of V1 cells in macaque monkey
Acquired by drifting sinusoidal gratings over receptive fields and measuring response
All acquired from the same location on retina

Cells respond preferential to specific spatial frequency bands - different but overlapping spatial frequency selectivities, they do Fourier analysis of the visual image

19
Q

What are David Marr’s 3 levels of analysis that must be applied in order to understand how a system works?

A

1) Computational level
What is the goal of the system?
i.e. what is purpose of vision - could be to recognise an object

2) Algorithmic level
What rules and representations can achieve this goal?

3) Implementational level
How is it achieved physically?

Vision serves multiple goals, including determining where objects are, what shape they have, and how to interact with them.

Each of these goals relies on numerous algorithmic steps.

20
Q

Edge detection is one of the algorithmic steps that helps us to determine what objects are and what shape they have.
Where do edges exist?

A

They only exist in interaction between observer and image
We focus on edges even when they are not inherently contained in an image

21
Q

What are the four steps in Marr’s model of object perception?

A

Gray-level - photoreceptors
Primal sketch - identifies object boundaries (edges)
2.5D sketch - depth perception
3D model

22
Q

What does Marr and Hildreth’s model of edge detection assume?

A

Assumes that edges of an object coincide with gradients in luminance
Only works if coincided luminance change where there is an edge
e.g. luminance gradient would seperate lake from tree

23
Q

What is involved in Marr and Hildreth’s (1980) model of edge detection - what are the steps? (first and second derivative)

A

Luminance gradient between two things
e.g. dark tree on light lake background

Take the first derivative of this - shows us sharp luminance changes that correspond to peaks/valleys where there is an intensity gradient
First derivative = rate of change in signal (across edge)

The second derivative is then taken, to stop signal being susceptible to noise - otherwise could be difficult to decide where peaks and valleys are

The second derivative gives us zero crossings (luminance changes across an edge - goes from negative to positive) where there is a luminance gradient
Edge not represented by peak or valley anymore

24
Q

What does Marr and Hildreth’s model allow computers to do?

A

Created algorithm so that computers can change image so only edges are highlighted

25
Why is there a smoothing process in edge detection?
This process will be negatively affected by high frequency noise - zero crossings where no meaningful gradient exists - detect edges where they are not actually present
26
What happens in the smoothing process of edge detection?
- Equivalent to first blurring the image (pre-derivatives) - Equivalent to removing high frequency content (i.e. analysis at coarse spatial scale) Expressed as convolving the image with a Gaussian operator G - each pixel is blurred with its neighbours - the sigma of G determines the level of blurring Higher level of sigma = more blurry image More blurry = more accurate edge detection, as edge detection is not corrupted by random noise
27
Does edge detection happen sequentially? i.e. Luminance change Smoothed First derivative Second derivative - zero crossing shows where edge is
No This can be done in all one step: Simple filtering operation - convolving the original image with a Laplacian of Gaussian filter achieves the same steps in one operation
28
How is the Laplacian of Gaussian filter implemented at a biological level?
2D model of it - rotationally invariant When this is plotted top-down It represents a retinal or LGN receptive field!
29
The smaller the LoG filter, the smaller the...
Spatial scale
30
What should LoG filters span?
A range of sizes so that the full range of scales and spatial frequencies are sampled There is a trade-off between noise removal (better at coarser scales) and edge enhancement (better at finer scales)
31
What is the spatial coincidence rule?
“If a zero-crossing segment is present in a set of independent channels (scales) over a continuous range of sizes… then the set of such zero crossing segments may be taken to indicate the presence of an intensity change in the image that is due to single physical phenomenon (a change in reflectance, illumination, depth or surface orientation)”. Marr and Hildreth
32
How do coarse and fine scales tell us if there is a meaningful edge?
Coarse scale = where edges are Finer scale = if edges in same place, edge information is meaningful Can use these to more accurately detect position of image Combining info from different spatial scales to get the trade off and best of both Retinal and LGN receptive fields are spatial filters that compute the second derivative of an image
33
How do off and on centre cells both inform simple cells in V1 of where an edge is?
Location of zero crossing is represented in both on and off centre cells Both cells excite simple cell - simple cell responds when both cells give a positive response Combine info to localise presence of edge - zero crossing
34
How does rapid edge detection (Paradiso and Nakayama (1991) show that low level visual perception is mostly edge based?
A target (white circle with black background) is presented for 16ms A mask (white ring on black background (smaller than white circle) - inner bit of circle is black) is then presented for 16ms Observer sees composite image of both target and mask Observers rate the brightness of the outside high and brightness of the inside dark Shows that observers see edges first, then rest of image is filled in - filling in process was interrupted by mask This shows that low level visual perception is mostly edge based
35
What are first order vs second order edges?
First order - defined by luminance gradient Second order - no luminance difference, defined by texture
36
What was Julesz's texton account of texture edge perception?
Julesz (1981) developed a model based on statistics of local conspicuous image features - textons Based on the simple idea that differences in the statistics of certain kinds of “conspicuous” features means certain features jump out at you - Oriented lines - Line terminations - Junctions (T and X junctions) Based on feature detection theory and preattentive visual search
37
What is the difference between effortless and difficult segmentation (Bergen and Julesz, 1983)
Effortless - difference in conspicuous element (e.g. line crossings) Difficult - no difference in conspicuous elements
38
How did Nothdurft (1985) critique texton account of texture edge perception?
Noticed similarities between luminance segmentation and texture segmentation Luminance segmentation becomes more difficult as element spacing increases - when textons are more spaced out, it is harder to detect a shape among the textons Even when features of textons are kept the same, segmentation ability can vary considerably Instead, these results suggest a mechanism that is more similar to an edge detection mechanism sensitive to spatial scale (i.e. spatial frequency) and orientation
39
What is Bergen and Adleson's (1988) Texture gradient theory of texture segmentation?
Evaluation of gradient (as for luminance) allows for texture segmentation Closer to Marr and Hildreth theory Making some textons longer = easier Making some textons shorter = harder - This is because it involves changing spatial frequency content of the image Different spatial scales = how we can do texture segmentation Computational models of texture segmentation (or edge detection) now explain this not in terms of textons but in terms of spatial differences in orientation and spatial frequency statistics This brings us back to “Fourier analysis”, in which an image is represented in terms of energy contained within channels tuned to combinations of spatial frequency and orientation
40
How does texture segmentation by texture gradient theory map onto neurophysiology? Lamme et al (1999)
Lamme et al (1999) used single cell recording to record from neurons in V1 in awake macaque monkey Receptive field was mapped and activity was recorded at different positions along the texture Temporal sequence of neural responses revealed separate processes (right figure shows fig – ground response): - Initial response (50 ms) - Local orientation tuning (58 ms) - Enhancement at edge (90 ms) - Enhancement in figure (112 - 123 ms) Consistent with an edge-based segmentation process that drives filling in of texture surface Edge enhancement at 90 ms suggests feedback from beyond V1. Only immediate sensitivity to the edge, with feedback from other visual areas Thus, low level vision is edge detection