Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 14 de 14
Filter
Add more filters










Publication year range
1.
PLoS Comput Biol ; 17(4): e1007907, 2021 04.
Article in English | MEDLINE | ID: mdl-33901165

ABSTRACT

The visual system is highly sensitive to spatial context for encoding luminance patterns. Context sensitivity inspired the proposal of many neural mechanisms for explaining the perception of luminance (brightness). Here we propose a novel computational model for estimating the brightness of many visual illusions. We hypothesize that many aspects of brightness can be explained by a dynamic filtering process that reduces the redundancy in edge representations on the one hand, while non-redundant activity is enhanced on the other. The dynamic filter is learned for each input image and implements context sensitivity. Dynamic filtering is applied to the responses of (model) complex cells in order to build a gain control map. The gain control map then acts on simple cell responses before they are used to create a brightness map via activity propagation. Our approach is successful in predicting many challenging visual illusions, including contrast effects, assimilation, and reverse contrast with the same set of model parameters.


Subject(s)
Contrast Sensitivity , Humans , Light , Models, Biological , Visual Perception
2.
J Comput Neurosci ; 48(1): 65-84, 2020 02.
Article in English | MEDLINE | ID: mdl-31980990

ABSTRACT

Hebbian plasticity means that if the firing of two neurons is correlated, then their connection is strengthened. Conversely, uncorrelated firing causes a decrease in synaptic strength. Spike-timing-dependent plasticity (STDP) represents one instantiation of Hebbian plasticity. Under STDP, synaptic changes depend on the relative timing of the pre- and post-synaptic firing. By inducing pre- and post-synaptic firing at different relative times the STDP curves of many neurons have been determined, and it has been found that there are different curves for different neuron types or synaptic sites. Biophysically, strengthening (long-term potentiation, LTP) or weakening (long-term depression, LTD) of glutamatergic synapses depends on the post-synaptic influx of calcium (Ca2+): weak influx leads to LTD, while strong, transient influx causes LTP. The voltage-dependent NMDA receptors are the main source of Ca2+ influx, but they will only open if a post-synaptic depolarisation coincides with pre-synaptic neurotransmitter release. Here we present a computational mechanism for Ca2+-dependent plasticity in which the interplay between the pre-synaptic neurotransmitter release and the post-synaptic membrane potential leads to distinct Ca2+ time-courses, which in turn lead to the change in synaptic strength. It is shown that the model complies with classic STDP results, as well as with results obtained with triplets of spikes. Furthermore, the model is capable of displaying different shapes of STDP curves, as observed in different experimental studies.


Subject(s)
Calcium Signaling/physiology , Models, Neurological , Neuronal Plasticity/physiology , Algorithms , Computer Simulation , Electrophysiological Phenomena/physiology , Glutamates/physiology , Humans , Long-Term Potentiation/physiology , Membrane Potentials/physiology , Neurons/physiology , Neurotransmitter Agents/metabolism , Receptors, AMPA/physiology , Receptors, N-Methyl-D-Aspartate/physiology
3.
Neural Netw ; 110: 66-81, 2019 Feb.
Article in English | MEDLINE | ID: mdl-30496916

ABSTRACT

The brain analyses the visual world through the luminance patterns that reach the retina. Formally, luminance (as measured by the retina) is the product of illumination and reflectance. Whereas illumination is highly variable, reflectance is a physical property that characterizes each object surface. Due to memory constraints, it seems plausible that the visual system suppresses illumination patterns before object recognition takes place. Since many combinations of reflectance and illumination can give rise to identical luminance values, finding the correct reflectance value of a surface is an ill-posed problem, and it is still an open question how it is solved by the brain. Here we propose a computational approach that first learns filter kernels ("receptive fields") for slow and fast variations in luminance, respectively, from achromatic real-world images. Distinguishing between luminance gradients (slow variations) and non-gradients (fast variations) could serve to constrain the mentioned ill-posed problem. The second stage of our approach successfully segregates luminance gradients and non-gradients from real-world images. Our approach furthermore predicts that visual illusions that contain luminance gradients (such as Adelson's checker-shadow display or grating induction) may occur as a consequence of this segregation process.


Subject(s)
Computer Simulation , Contrast Sensitivity , Cues , Lighting/methods , Pattern Recognition, Visual , Photic Stimulation/methods , Computer Simulation/trends , Contrast Sensitivity/physiology , Humans , Pattern Recognition, Visual/physiology , Visual Perception/physiology
4.
PLoS Comput Biol ; 11(10): e1004479, 2015 Oct.
Article in English | MEDLINE | ID: mdl-26513150

ABSTRACT

Power laws describe brain functions at many levels (from biophysics to psychophysics). It is therefore possible that they are generated by similar underlying mechanisms. Previously, the response properties of a collision-sensitive neuron were reproduced by a model which used a power law for scaling its inhibitory input. A common characteristic of such neurons is that they integrate information across a large part of the visual field. Here we present a biophysically plausible model of collision-sensitive neurons with η-like response properties, in which we assume that each information channel is noisy and has a response threshold. Then, an approximative power law is obtained as a result of pooling these channels. We show that with this mechanism one can successfully predict many response characteristics of the Lobula Giant Movement Detector Neuron (LGMD). Moreover, the results depend critically on noise in the inhibitory pathway, but they are fairly robust against noise in the excitatory pathway.


Subject(s)
Dendrites/physiology , Models, Neurological , Models, Statistical , Motion Perception/physiology , Neurons, Afferent/physiology , Synapses/physiology , Animals , Computer Simulation , Differential Threshold/physiology , Grasshoppers/physiology , Sensory Thresholds/physiology , Signal-To-Noise Ratio
5.
Vision Res ; 92: 53-8, 2013 Nov.
Article in English | MEDLINE | ID: mdl-24075899

ABSTRACT

The different sources of information that can be used to estimate time-to-contact may have different degrees of reliability across time. For example, after a given presentation or display time, an absolute change of angular size can be more reliable than the corresponding estimation of the rate of angular expansion (e.g. motion information). One could then expect systematic biases in the observer's responses for different times of stimulus exposure. In one experiment, observers judged whether approaching objects arrived at the point of observation before or after a reference beep (1.2s) under monocular, and binocular plus monocular vision. Five display times from 0.1 to 0.9s were used. Unlike monocular viewing, where accuracy increased monotonically with display time, an interesting non-linearity occurred for objects with small size when binocular information was available. Accuracy reached maximum values for small objects with only 0.3s of vision with stereopsis. This accuracy, however, dropped significantly after 0.4s of exposure and increased again linearly with time. This is consistent with subjects switching from using binocular information to using monocular motion information when it started to become more reliable. We also explored whether monocular cues were combined differently across time by fitting a model that relates visual angle to its rate of expansion. Results show that subjects relied more on angular motion information (i.e. rate of expansion) with presentation time but interrupting this motion integration process led to a loss of accuracy in time-to-contact judgments.


Subject(s)
Depth Perception/physiology , Motion Perception/physiology , Time Perception/physiology , Adult , Humans , Photic Stimulation/methods , Psychophysics , Vision, Binocular/physiology , Vision, Monocular/physiology
6.
PLoS Comput Biol ; 8(8): e1002625, 2012.
Article in English | MEDLINE | ID: mdl-22915999

ABSTRACT

The τ-function and the η-function are phenomenological models that are widely used in the context of timing interceptive actions and collision avoidance, respectively. Both models were previously considered to be unrelated to each other: τ is a decreasing function that provides an estimation of time-to-contact (ttc) in the early phase of an object approach; in contrast, g has a maximum before ttc. Furthermore, it is not clear how both functions could be implemented at the neuronal level in a biophysically plausible fashion. Here we propose a new framework--the corrected modified Tau function--capable of predicting both τ-type ("τ(cm)") and g-type ("t(mod)") responses. The outstanding property of our new framework is its resilience to noise. We show that t(mod) can be derived from a firing rate equation, and, as g, serves to describe the response curves of collision sensitive neurons. Furthermore, we show that tcm predicts the psychophysical performance of subjects determining ttc. Our new framework is thus validated successfully against published and novel experimental data. Within the framework, links between τ-type and η-type neurons are established. Therefore, it could possibly serve as a model for explaining the co-occurrence of such neurons in the brain.


Subject(s)
Neurons/physiology , Animals , Humans , Models, Theoretical , Psychophysics , Species Specificity , tau Proteins/physiology
7.
PLoS One ; 7(4): e35705, 2012.
Article in English | MEDLINE | ID: mdl-22558205

ABSTRACT

The visual angle that is projected by an object (e.g. a ball) on the retina depends on the object's size and distance. Without further information, however, the visual angle is ambiguous with respect to size and distance, because equal visual angles can be obtained from a big ball at a longer distance and a smaller one at a correspondingly shorter distance. Failure to recover the true 3D structure of the object (e.g. a ball's physical size) causing the ambiguous retinal image can lead to a timing error when catching the ball. Two opposing views are currently prevailing on how people resolve this ambiguity when estimating time to contact. One explanation challenges any inference about what causes the retinal image (i.e. the necessity to recover this 3D structure), and instead favors a direct analysis of optic flow. In contrast, the second view suggests that action timing could be rather based on obtaining an estimate of the 3D structure of the scene. With the latter, systematic errors will be predicted if our inference of the 3D structure fails to reveal the underlying cause of the retinal image. Here we show that hand closure in catching virtual balls is triggered by visual angle, using an assumption of a constant ball size. As a consequence of this assumption, hand closure starts when the ball is at similar distance across trials. From that distance on, the remaining arrival time, therefore, depends on ball's speed. In order to time the catch successfully, closing time was coupled with ball's speed during the motor phase. This strategy led to an increased precision in catching but at the cost of committing systematic errors.


Subject(s)
Motion Perception/physiology , Pattern Recognition, Visual/physiology , Psychomotor Performance/physiology , Hand/physiology , Humans , Motion , Optic Flow/physiology , Task Performance and Analysis , Time Factors , Visual Fields/physiology
8.
PLoS Comput Biol ; 5(3): e1000329, 2009 Mar.
Article in English | MEDLINE | ID: mdl-19325870

ABSTRACT

Numerous psychophysical experiments found that humans preferably rely on a narrow band of spatial frequencies for recognition of face identity. A recently conducted theoretical study by the author suggests that this frequency preference reflects an adaptation of the brain's face processing machinery to this specific stimulus class (i.e., faces). The purpose of the present study is to examine this property in greater detail and to specifically elucidate the implication of internal face features (i.e., eyes, mouth, and nose). To this end, I parameterized Gabor filters to match the spatial receptive field of contrast sensitive neurons in the primary visual cortex (simple and complex cells). Filter responses to a large number of face images were computed, aligned for internal face features, and response-equalized ("whitened"). The results demonstrate that the frequency preference is caused by internal face features. Thus, the psychophysically observed human frequency bias for face processing seems to be specifically caused by the intrinsic spatial frequency content of internal face features.


Subject(s)
Face/anatomy & histology , Models, Biological , Models, Neurological , Nerve Net/physiology , Pattern Recognition, Visual/physiology , Visual Cortex/physiology , Biomimetics/methods , Computer Simulation , Female , Humans , Male
9.
PLoS One ; 3(7): e2590, 2008 Jul 02.
Article in English | MEDLINE | ID: mdl-18596932

ABSTRACT

Psychophysical studies suggest that humans preferentially use a narrow band of low spatial frequencies for face recognition. Here we asked whether artificial face recognition systems have an improved recognition performance at the same spatial frequencies as humans. To this end, we estimated recognition performance over a large database of face images by computing three discriminability measures: Fisher Linear Discriminant Analysis, Non-Parametric Discriminant Analysis, and Mutual Information. In order to address frequency dependence, discriminabilities were measured as a function of (filtered) image size. All three measures revealed a maximum at the same image sizes, where the spatial frequency content corresponds to the psychophysical found frequencies. Our results therefore support the notion that the critical band of spatial frequencies for face recognition in humans and machines follows from inherent properties of face images, and that the use of these frequencies is associated with optimal face recognition performance.


Subject(s)
Face , Image Interpretation, Computer-Assisted/methods , Pattern Recognition, Automated/methods , Female , Humans , Image Interpretation, Computer-Assisted/instrumentation , Male , Pattern Recognition, Visual , Visual Perception
10.
Proc Biol Sci ; 275(1647): 2095-100, 2008 Sep 22.
Article in English | MEDLINE | ID: mdl-18544506

ABSTRACT

Psychophysical experiments suggested a relative importance of a narrow band of spatial frequencies for recognition of face identity in humans. There exists, however, no conclusive evidence of why it is that such frequencies are preferred. To address this question, I examined the amplitude spectra of a large number of face images and observed that face spectra generally fall off more steeply with spatial frequency compared with ordinary natural images. When external face features (such as hair) are suppressed, then whitening of the corresponding mean amplitude spectra revealed higher response amplitudes at those spatial frequencies which are deemed important for processing face identity. The results presented here therefore provide support for that face processing characteristics match corresponding stimulus properties.


Subject(s)
Face/anatomy & histology , Pattern Recognition, Visual/physiology , Female , Humans , Image Processing, Computer-Assisted , Male
11.
Vision Res ; 47(27): 3360-72, 2007 Dec.
Article in English | MEDLINE | ID: mdl-17998141

ABSTRACT

The neuronal mechanisms that serve to distinguish between light emitting and light reflecting objects are largely unknown. It has been suggested that luminosity perception implements a separate pathway in the visual system, such that luminosity constitutes an independent perceptual feature. Recently, a psychophysical study was conducted to address the question whether luminosity has a feature status or not. However, the results of this study lend support to the hypothesis that luminance gradients are instead a perceptual feature. Here, I show how the perception of luminosity can emerge from a previously proposed neuronal architecture for generating representations of luminance gradients.


Subject(s)
Models, Neurological , Models, Psychological , Visual Perception/physiology , Glare , Humans , Lighting , Psychophysics
12.
Vision Res ; 46(17): 2659-74, 2006 Sep.
Article in English | MEDLINE | ID: mdl-16603218

ABSTRACT

Recent evidence suggests that object surfaces and their properties are represented at early stages in the visual system of primates. Most likely invariant surface properties are extracted to endow primates with robust object recognition capabilities. In real-world scenes, luminance gradients are often superimposed on surfaces. We argue that gradients should also be represented in the visual system, since they encode highly variable information, such as shading, focal blur, and penumbral blur. We present a neuronal architecture which was designed and optimized for segregating and representing luminance gradients in real-world images. Our architecture in addition provides a novel theory for Mach bands, whereby corresponding psychophysical data are predicted consistently.


Subject(s)
Contrast Sensitivity/physiology , Models, Neurological , Models, Psychological , Pattern Recognition, Visual/physiology , Humans , Lighting , Psychophysics , Retinal Ganglion Cells/physiology , Sensory Thresholds/physiology , Surface Properties
13.
Neural Comput ; 18(4): 871-903, 2006 Apr.
Article in English | MEDLINE | ID: mdl-16494694

ABSTRACT

Recent evidence suggests that the primate visual system generates representations for object surfaces (where we consider representations for the surface attribute brightness). Object recognition can be expected to perform robustly if those representations are invariant despite environmental changes (e.g., in illumination). In real-world scenes, it happens, however, that surfaces are often overlaid by luminance gradients, which we define as smooth variations in intensity. Luminance gradients encode highly variable information, which may represent surface properties (curvature), nonsurface properties (e.g., specular highlights, cast shadows, illumination inhomogeneities), or information about depth relationships (cast shadows, blur). We argue, on grounds of the unpredictable nature of luminance gradients, that the visual system should establish corresponding representations, in addition to surface representations. We accordingly present a neuronal architecture, the so-called gradient system, which clarifies how spatially accurate gradient representations can be obtained by relying on only high-resolution retinal responses. Although the gradient system was designed and optimized for segregating, and generating, representations of luminance gradients with real-world luminance images, it is capable of quantitatively predicting psychophysical data on both Mach bands and Chevreul's illusion. It furthermore accounts qualitatively for a modified Ehrenstein disk.


Subject(s)
Contrast Sensitivity/physiology , Models, Neurological , Pattern Recognition, Visual/physiology , Animals , Geniculate Bodies/physiology , Luminescence , Retina/physiology
14.
Neural Netw ; 18(10): 1319-31, 2005 Dec.
Article in English | MEDLINE | ID: mdl-16039097

ABSTRACT

Filling-in models were successful in predicting psychophysical data for brightness perception. Nevertheless, their suitability for real-world image processing has never been examined. A unified architecture for both predicting psychophysical data and real-world image processing would constitute a powerful theory for early visual information processing. As a first contribution of the present paper, we identified three principal problems with current filling-in architectures, which hamper the goal of having such a unified architecture. To overcome these problems we propose an advance to filling-in theory, called BEATS filling-in, which is based on a novel nonlinear diffusion operator. BEATS filling-in furthermore introduces novel boundary structures. We compare, by means of simulation studies with real-world images, the performance of BEATS filling-in with the recently proposed confidence-based filling-in. As a second contribution we propose a novel mechanism for encoding luminance information in contrast responses ('multiplex contrasts'), which is based on recent neurophysiological findings. Again, by simulations, we show that 'multiplex contrasts' at a single, high-resolution filter scale are sufficient for recovering absolute luminance levels. Hence, 'multiplex contrasts' represent a novel theory addressing how the brain encodes and decodes luminance information.


Subject(s)
Computer Systems , Image Processing, Computer-Assisted , Computer Simulation , Humans , Light , Models, Neurological , Models, Statistical , Nonlinear Dynamics , Psychophysics , Retina/physiology , Visual Perception
SELECTION OF CITATIONS
SEARCH DETAIL
...