scholarly journals On designing a visual system (Towards a Gibsonian computational model of vision)

1989 ◽  
Vol 1 (4) ◽  
pp. 289-337 ◽  
Author(s):  
AARON SLOMAN
2020 ◽  
Author(s):  
Alejandro Lerer ◽  
Hans Supèr ◽  
Matthias S.Keil

AbstractThe visual system is highly sensitive to spatial context for encoding luminance patterns. Context sensitivity inspired the proposal of many neural mechanisms for explaining the perception of luminance (brightness). Here we propose a novel computational model for estimating the brightness of many visual illusions. We hypothesize that many aspects of brightness can be explained by a predictive coding mechanism, which reduces the redundancy in edge representations on the one hand, while non-redundant activity is enhanced on the other (response equalization). Response equalization is implemented with a dynamic filtering process, which (dynamically) adapts to each input image. Dynamic filtering is applied to the responses of complex cells in order to build a gain control map. The gain control map then acts on simple cell responses before they are used to create a brightness map via activity propagation. Our approach is successful in predicting many challenging visual illusions, including contrast effects, assimilation, and reverse contrast.Author summaryWe hardly notice that what we see is often different from the physical world “outside” of the brain. This means that the visual experience that the brain actively constructs may be different from the actual physical properties of objects in the world. In this work, we propose a hypothesis about how the visual system of the brain may construct a representation for achromatic images. Since this process is not unambiguous, sometimes we notice “errors” in our perception, which cause visual illusions. The challenge for theorists, therefore, is to propose computational principles that recreate a large number of visual illusions and to explain why they occur. Notably, our proposed mechanism explains a broader set of visual illusions than any previously published proposal. We achieved this by trying to suppress predictable information. For example, if an image contained repetitive structures, then these structures are predictable and would be suppressed. In this way, non-predictable structures stand out. Predictive coding mechanisms act as early as in the retina (which enhances luminance changes but suppresses uniform regions of luminance), and our computational model holds that this principle also acts at the next stage in the visual system, where representations of perceived luminance (brightness) are created.


2012 ◽  
Vol 8 (4) ◽  
pp. 387-415
Author(s):  
Marc Ebner

ABSTRACT Color is not a physical quantity of an object. It cannot be measured. We can only measure reflectance, i.e. the amount of light reflected for each wavelength. Nevertheless, we attach colors to the objects around us. A human observer perceives colors as being approximately constant irrespective of the illuminant which is used to illuminate the scene. Colors are a very important cue in everyday life. They can be used to recognize or distinguish different objects. Currently, we do not yet know how the brain arrives at a color constant or approximately color constant descriptor, i.e. what computational processing is actually performed by the brain. What we need is a computational description of color perception in particular and color vision in general. Only if we are able to write down a full computational theory of the visual system then we have understood how the visual system works. With this contribution, a computational model of color perception is presented. This model is much simpler compared to previous theories. It is able to compute a color constant descriptor even in the presence of spatially varying illuminants. According to this model, the cones respond approximately logarithmic to the irradiance entering the eye. Cells in V1 perform a change of the coordinate system such that colors are represented along a red-green, a blue-yellow and a black-white axis. Cells in V4 compute local space average color using a resistive grid. The resistive grid is formed by cells in V4. The left and right hemispheres are connected via the corpus callosum. A color constant descriptor which is presumably used for color based object recognition is computed by subtracting local space average color from the cone response within a rotated coordinate system.


2002 ◽  
Vol 44-46 ◽  
pp. 881-887 ◽  
Author(s):  
T. Koshizen ◽  
Koji Akatsuka ◽  
Hiroshi Tsujino

2016 ◽  
Vol 371 (1697) ◽  
pp. 20150258 ◽  
Author(s):  
P. Cammack ◽  
J. M. Harris

Deciding what constitutes an object, and what background, is an essential task for the visual system. This presents a conundrum: averaging over the visual scene is required to obtain a precise signal for object segregation, but segregation is required to define the region over which averaging should take place. Depth, obtained via binocular disparity (the differences between two eyes’ views), could help with segregation by enabling identification of object and background via differences in depth. Here, we explore depth perception in disparity-defined objects. We show that a simple object segregation rule, followed by averaging over that segregated area, can account for depth estimation errors. To do this, we compared objects with smoothly varying depth edges to those with sharp depth edges, and found that perceived peak depth was reduced for the former. A computational model used a rule based on object shape to segregate and average over a central portion of the object, and was able to emulate the reduction in perceived depth. We also demonstrated that the segregated area is not predefined but is dependent on the object shape. We discuss how this segregation strategy could be employed by animals seeking to deter binocular predators. This article is part of the themed issue ‘Vision in our three-dimensional world’.


Sign in / Sign up

Export Citation Format

Share Document