From receptor signal to perception Flashcards
what do you look at and what do you see?
- Looking: photoreceptor signals change v. quickly and are noisy
○ Movements (eyes, head and body) change and stabilise gaze for very shirt periods of time, fast, main function filtering, decomposing images into elementary features in peripheral layers○ of the visual system, early segregation of signals fed into parallel visual streams for unconscious and conscious visual perception ○ Very little to do with what we consciously perceive 2. Actively looking and seeing: unconscious perception can be fast or slow, filtered depending on task (pathway), can be invariant, selective, is less noisy ○ Guides fast and slow actions, analyses scenes and object, changes are perceived through task outcomes (e.g. consequences or a behavioural response, change in internal state or top-down control (e.g. gaze, conscious decision-making in humans)) 3. Actively seeing: conscious visual perception in humans is stable, slow, invariant, affected by filtering but less selective than unconscious perception, low noise ○ Only some pronounced eye, head or body movements result in a perceived change of the viewed scene or object, conscious vision seems relevant for some specific tasks and can exert some top-down control but mostly results from processes at level 1 and 2, some of which are hard-wired • Humans: difficult to generate evidence that clearly separates domains 2 and 3 • Animals: level 3 not assumed Try to separate conscious and unconscious perception to see what special tasks conscious vision has to accomplish
Early stages of visual processing involve edge filtering and enhancement
• LI (found in the CS RFs of many ganglion cells in the retina) can explain hard-wired phenom in perception, such as the Herrman grid
• We see grey dots that are not there
• During eye movements the same part of the image is viewed by foeval receptors and ganglion cells or peripheral ones
• The CS RFs are smallest in the fovea (highest spatial resolution) and larger the further in the periphery of the retina they are (dots appear grey) - dominated by black tiles
• Move fovea to where you want to explore - dots become white
• Subtract signals of centre from surround
• Vary size of squares and get same effect - if completely hard wired to retinal ganglion cells, shouldn’t be happening - probably further processes at level of LGN and cortex - other types of receptive fields - modern criticism
Looking at grid doesn’t generate any useful behaviour - criticism - artificial scenes/patterns
see notes
Zanker (2010)
- Mach bands (optical illusion described by physicist Ernst Mach in 1865)
see notes
* From left to right - increasing brightness * Exaggerated towards darker or brighter area • High contrast edges • Removes illusion But can change back - hard wired into perception
see notes
• On-centre OFF-surround example • Ganglion cells with CS RFs: ○ Enhanced edges ○ Compresses information (only respond when in RF) ○ Filters info according to spatial frequencies (diff sized RFs and varying sensitivity across retina) • Boundary falls somewhere on retina • Direct fovea • Eyes move and have number of signals that vary over time • Number of CS RFs processing edge • CS RF don’t respond/change signal when illumination uniform • When illumination changes get either inhib or excitatory signal • Ganglion cells have average resting freq • Delta - change in frequency • 0 = no change in freq Resting state it still fires
see notes
• Thinner striped - higher spatial freq - can fit more stripes • One way to characterise the quality of vision Refined perceptual test
Early stages of visual processing involve edge filtering and enhancement research
Bakshi and Ghosh (2020)
Skottun (2020)
Khosravy et al. (2017)
Ghosh et al. (2006)
Misson and Anderson (2017)
De Valois et al. (2014)
Bakshi and Ghosh (2020)
A novel modification of the Hermann grid stimulus is demonstrated. It is shown that introduction of extremely tiny squares at the corners of the grid squares in the classical stimulus, keeping the position and orientation of the grid squares fixed, can reduce the strength and even completely wipe out the illusory dark spots. The novel perturbing stimulus was investigated further and a gray-level intensity threshold was measured for the tiny corner squares beyond which the illusory blobs disappear completely. It was also found that this threshold remains practically unchanged over a wide range of grid square size for an observer.
Skottun (2020)
The Hermann Grid is made up of a series of vertical and horizontal bars. The Hermann Grid Illusion consists in the brightness of the intersections appearing different from that of the sections between intersections in spite of the luminance being the same. In the case of a light grid on a dark background the intersections tend to appear darker than the parts between intersections. It is here pointed out, in two different ways, that the stimulus power is less for the parts of the grid located at intersections than for parts of the grid between intersections. This is all in the stimuli and does not depend on vision or the visual system. Were we to assume that a stronger stimulus gives a brighter appearance this would make the parts between intersections appear brighter than the parts of the grid at intersections. This would be consistent with the Hermann Grid Illusion.
Khosravy et al. (2017)
The perceptual adaptation of the image (PAI) is introduced by inspiration from Chevreul-Mach Bands (CMB) visual phenomenon. By boosting the CMB assisting illusory effect on boundaries of the regions, PAI adapts the image to the perception of the human visual system and thereof increases the quality of the image. PAI is proposed for application to standard images or the output of any image processing technique. For the implementation of the PAI on the image, an algorithm of morphological filters (MFs) is presented, which geometrically adds the model of CMB effect. Numerical evaluation by improvement ratios of four no-reference image quality assessment (NR-IQA) indexes approves PAI performance where it can be noticeably observed in visual comparisons. Furthermore, PAI is applied as a postprocessing block for classical morphological filtering, weighted morphological filtering, and median morphological filtering in cancelation of salt and pepper, Gaussian, and speckle noise from MRI images, where the above specified NR-IQA indexes validate it. PAI effect on image enhancement is benchmarked upon morphological image sharpening and high-boost filtering.
Ghosh et al. (2006)
A re-scan of the well-known Mach band illusion has led to the proposal of a Bi-Laplacian of Gaussian operation in early vision. Based on this postulate, the human visual system at low-level has been modeled from two approaches that give rise to two new tools. On one hand, it leads to the construction of a new image sharpening kernel, and on the other, to the explanation of more complex brightness-contrast illusions and the possible development of a new algorithm for robust visual capturing and display systems.
Misson and Anderson (2017)
It is generally believed that humans perceive linear polarized light following its conversion into a luminance signal by diattenuating macular structures. Measures of polarization sensitivity may therefore allow a targeted assessment of macular function. Our aim here was to quantify psychophysical characteristics of human polarization perception using grating and optotype stimuli defined solely by their state of linear polarization. We show: (i) sensitivity to polarization patterns follows the spectral sensitivity of macular pigment; (ii) the change in sensitivity across the central field follows macular pigment density; (iii) polarization patterns are identifiable across a range of contrasts and scales, and can be resolved with an acuity of 15.4 cycles/degree (0.29 logMAR); and (iv) the human eye can discriminate between areas of linear polarization differing in electric field vector orientation by as little as 4.4 degrees. These findings, which support the macular diattenuator model of polarization sensitivity, are unique for vertebrates and approach those of some invertebrates with a well-developed polarization sense. We conclude that this sensory modality extends beyond Haidinger’s brushes to the recognition of quantifiable spatial polarization-modulated patterns. Furthermore, the macular origin and sensitivity of human polarization pattern perception makes it potentially suitable for the detection and quantification of macular dysfunction.
De Valois et al. (2014)
The detectability of luminance modulated gratings of different spatial frequencies was determined at five different adaptation levels for three macaque monkeys and five normal human observers. The human and macaque observers gave results which were identical in form and similar in absolute values. Both species showed optimal contrast sensitivity in the middle spatial frequency range of about 3–5 c/deg with both low and high frequency attenuation, at high light levels. Contrast sensitivity to high frequencies dropped rapidly as adaptation levels were lowered, with a resulting shift in peak sensitivity to lower spatial frequencies. At the lowest adaptation level studied, neither macaque nor human observers showed any low frequency attenuation in the spatial luminance contrast sensitivity function.
Filtering of images and scenes at different spatial frequencies
see notes
• Filter different spatial frequencies • Place diff sized stripes over RF in fovea get number of diff signals and ratio between white and black changes • One way neurons code with CS RFs to filter out diff signals at high and low spatial freqs • Changing frequency useful for recognising emotions (Mona Lisa image) • Neurons that record intensity of high spatial freq components in image useful to recognise particular features, to know what features are and what small diffs are in facial expression • Takes lots of processing power • Time crucial factor to determine whether certain brain systems will use visual input that is filtered through low or high spatial freq filters • Not mutually exclusive • Evidence at neuronal level for both At first stage, ganglion cells with CS RFs and then more complicated as look at more central brain layers
Filtering of images and scenes at different spatial frequencies research
Estevez et al. (2016)
Logunova and Shelepina (2015)
Estevez et al. (2016)
The use of apodizing or superresolving filters improves the performance of an optical system in different frequency bands. This improvement can be seen as an increase in the OTF value compared to the OTF for the clear aperture.
In this paper we propose a method to enhance the contrast of an image in both its low and its high frequencies. The method is based on the generation of a synthetic Optical Transfer Function, by multi-plexing the OTFs given by the use of different non-uniform transmission filters on the pupil. We propose to capture three images, one obtained with a clear pupil, one obtained with an apodizing filter that enhances the low frequencies and another one taken with a superresolving filter that improves the high frequencies. In the Fourier domain the three spectra are combined by using smoothed passband filters, and then the inverse transform is performed. We show that we can create an enhanced image better than the image obtained with the clear aperture. To evaluate the performance of the method, bar tests (sinusoidal tests) with different frequency content are used. The results show that a contrast improvement in the high and low frequencies is obtained
Logunova and Shelepina (2015)
This paper discusses the process of interpreting scenes with the image of a human face, subjected to processing with spatial-frequency filters that simulate the characteristics of the receptive fields of the neurons of the primary visual cortex. A technique was used that makes it possible to give a quantitative evaluation of the interpretation of an image while carrying out tasks of identifying a period of emotional stress and the age-related features of the person. It was shown that, besides the horizontal components of the spatial-frequency spectrum, a substantial role is played in the process of interpreting images of faces by the diagonal components. Even though the visual system is less sensitive to the diagonal components than to the horizontal ones, the information contained in them makes it possible not only to distinguish age-related features, but also to give the supplementary information needed to identify an unfamiliar person when encountering that person again
Inter- and intraspecific variations in acuity and contrast sensitivity
(Owsley, 2016)
see notes
• Ghim and Hodos (2006) ○ Range of diff filter functions that differ between indvs Make inferences about how the world will look diff to diff indvs/species
see notes
Inter- and intraspecific variations in acuity and contrast sensitivity research
Gruber et al. (2013)
Billino and Pilz (2019)
Potier et al. (2018)
Feng et al. (2017)
Gruber et al. (2013)
Objective:In this article, we review the impact of vision on older people’s night driving abilities. Driving is the preferred and primary mode of transport for older people. It is a complex activity where intact vision is seminal for road safety. Night driving requires mesopic rather than scotopic vision, because there is always some light available when driving at night.Scotopicrefers to night vision,photopicrefers to vision under well-lit conditions, andmesopicvision is a combination of photopic and scotopic vision in low but not quite dark lighting situations. With increasing age, mesopic vision decreases and glare sensitivity increases, even in the absence of ocular diseases. Because of the increasing number of elderly drivers, more drivers are affected by night vision difficulties. Vision tests, which accurately predict night driving ability, are therefore of great interest.
Methods:We reviewed existing literature on age-related influences on vision and vision tests that correlate or predict night driving ability.
Results:We identified several studies that investigated the relationship between vision tests and night driving. These studies found correlations between impaired mesopic vision or increased glare sensitivity and impaired night driving, but no correlation was found among other tests; for example, useful field of view or visual field. The correlation between photopic visual acuity, the most commonly used test when assessing elderly drivers, and night driving ability has not yet been fully clarified.
Conclusions:Photopic visual acuity alone is not a good predictor of night driving ability. Mesopic visual acuity and glare sensitivity seem relevant for night driving. Due to the small number of studies evaluating predictors for night driving ability, further research is needed.
Billino and Pilz (2019)
Research on functional changes across the adult lifespan has been dominated by studies related to cognitive processes. However, it has become evident that a more comprehensive approach to behavioral aging is needed. In particular, our understanding of age-related perceptual changes is limited. Visual motion perception is one of the most studied areas in perceptual aging and therefore, provides an excellent domain on the basis of which we can investigate the complexity of the aging process. We review the existing literature on how aging affects motion perception, including different processing stages, and consider links to cognitive and motor changes. We address the heterogeneity of results and emphasize the role of individual differences. Findings on age-related changes in motion perception ultimately illustrate the complexity of functional dynamics that can contribute to decline as well as stability during healthy aging. We thus propose that motion perception offers a conceptual framework for perceptual aging, encouraging a deliberate consideration of functional limits and resources emerging across the lifespan.
Potier et al. (2018)
Animals are thought to use achromatic signals to detect small (or distant) objects and chromatic signals for large (or nearby) objects. While the spatial resolution of the achromatic channel has been widely studied, the spatial resolution of the chromatic channel has rarely been estimated. Using an operant conditioning method, we determined (i) the achromatic contrast sensitivity function and (ii) the spatial resolution of the chromatic channel of a diurnal raptor, the Harris’s hawkParabuteo unicinctus. The maximal spatial resolution for achromatic gratings was 62.3 c deg−1, but the contrast sensitivity was relatively low (10.8–12.7). The spatial resolution for isoluminant red-green gratings was 21.6 c deg−1—lower than that of the achromatic channel, but the highest found in the animal kingdom to date. Our study reveals that Harris’s hawks have high spatial resolving power for both achromatic and chromatic vision, suggesting the importance of colour vision for foraging. By contrast, similar to other bird species, Harris’s hawks have low contrast sensitivity possibly suggesting a trade-off with chromatic sensitivity. The result is interesting in the light of the recent finding that double cones—thought to mediate high-resolution vision in birds—are absent in the central fovea of raptors.
Feng et al. (2017)
In humans, geometrical illusions are thought to reflect mechanisms that are usually helpful for seeing the world in a predictable manner. These mechanisms deceive us given the right set of circumstances, correcting visual input where a correction is not necessary. Investigations of non-human animals’susceptibility to geometrical illusions have yielded contradictory results, suggesting that the underlying mechanisms with which animals see the world may differ across species. In this review, we first collate studies showing that different species are susceptible to specific illusions in the same or reverse direction as humans. Based on a careful assessment of these findings, we then propose several ecological and anatomical factors that may affect how a species perceives illusory stimuli. We also consider the usefulness of this information for determining whether sight in different species might be more similar to human sight, being influenced by contextual information, or to how machines process and transmit information as programmed. Future testing in animals could provide new theoretical insights by focusing on establishing dissociations between stimuli that may or may not alter perception in a particular species. This information could improve our understanding of the mechanisms behind illusions, but also provide insight into how sight is subjectively experienced by different animals, and the degree to which vision is innate versus acquired, which is difficult to examine in humans
can you find the badger?
• Seeing and recognising objects, mates, predators or prey is imp for many tasks
• But visual scenes often crowded (and typically not black and white)
• Contrast enhancement of edges is imp for many visual tasks - objects characterised by their edges
• A major task of the visual system is to segregate objects and backgrounds, automatically and quickly - based on analysis of edges - how fast they move - motion information - happens automatically and quickly - may not be able to influence it easily or at all
Other tasks require further computations in order to extract info - e.g. face recognition task
Insect lands preferably at the edge of objects (Eglehaaf et al., 2012; Kang et al., 2012)
• Moths actively choose spot and vary their orientation to align with the lines in the background for better camouflage against avian predators
• Recording natural landing behav of fly on cup - requires lots of coord and body posture control - controlling speed
Land at contrast edges - boundaries of objects
see notes
After landing can reposition - main orientation on bark that is signalled by contrast edges
see notes
Insect lands preferably at the edge of objects (Eglehaaf et al., 2012; Kang et al., 2012) research
Egelhaaf et al. (2014)
Mauss and Borst (2020)
Kang et al. (2015)
Green et al. (2019)
Egelhaaf et al. (2014)
Despite their miniature brains insects, such as flies, bees and wasps, are able to navigate by highly erobatic flight maneuvers in cluttered environments. They rely on spatial information that is contained in the retinal motion patterns induced on the eyes while moving around (“optic flow”) to accomplish their extraordinary performance. Thereby, they employ an active flight and gaze strategy that separates rapid saccade-like turns from translatory flight phases where the gaze direction is kept largely constant. This behavioral strategy facilitates the processing of environmental information, because information about the distance of the animal to objects in the environment is only contained in the optic flow generated by translatory motion. However, motion detectors as are widespread in biological systems do not represent veridically the velocity of the optic flow vectors, but also reflect textural information about the environment. This characteristic has often been regarded as a limitation of a biological motion detection mechanism. In contrast, we conclude from analyses challenging insect movement detectors with image flow as generated during translatory locomotion through cluttered natural environments that this mechanism represents the contours of nearby objects. Contrast borders are a main carrier of functionally relevant object information in artificial and natural sceneries. The motion detection system thus segregates in a computationally parsimonious way the environment into behaviorally relevant nearby objects and—in many behavioral contexts—less relevant distant structures. Hence, by making use of an active flight and gaze strategy, insects are capable of performing extraordinarily well even with a computationally simple motion detection mechanism.
Mauss and Borst (2020)
○ Optic flow arising from self-motion provides a rich source of information.
○ Optic flow detection and related behaviors have been studied extensively in insects.
○ Translational flow affordsspatial visionand estimation of travel speed.
○ Rotational flow mediates estimation and compensation of involuntary course changes
○ All optic flow-based behaviors likely depend on the same local motion detectors.
Vision is an important sensory modality for navigation in roaming animals. In contrast to most vertebrates, insects usually must cope with low resolution retinal images and the inability to infer spatial features using accommodation or stereovision. However, during locomotion, the retinal input is dominated by characteristic panoramic image shifts, termed optic flow, that depend on self-motion parameters and environmental features. Therefore, optic flow provides a rich source of information guiding locomotion speed as well as the position and orientation of animals over time relative to their surroundings. Here, focusing on flight behavior, we describe the strategies and putative underlying neuronal mechanisms by which insects control their course through processing of visual motion cues.