Multisensory Bayesian Inference Depends on Synapse Maturation during Training: Theoretical Analysis and Neural Modeling Implementation

2017 ◽  
Vol 29 (3) ◽  
pp. 735-782 ◽  
Author(s):  
Mauro Ursino ◽  
Cristiano Cuppini ◽  
Elisa Magosso

Recent theoretical and experimental studies suggest that in multisensory conditions, the brain performs a near-optimal Bayesian estimate of external events, giving more weight to the more reliable stimuli. However, the neural mechanisms responsible for this behavior, and its progressive maturation in a multisensory environment, are still insufficiently understood. The aim of this letter is to analyze this problem with a neural network model of audiovisual integration, based on probabilistic population coding—the idea that a population of neurons can encode probability functions to perform Bayesian inference. The model consists of two chains of unisensory neurons (auditory and visual) topologically organized. They receive the corresponding input through a plastic receptive field and reciprocally exchange plastic cross-modal synapses, which encode the spatial co-occurrence of visual-auditory inputs. A third chain of multisensory neurons performs a simple sum of auditory and visual excitations. The work includes a theoretical part and a computer simulation study. We show how a simple rule for synapse learning (consisting of Hebbian reinforcement and a decay term) can be used during training to shrink the receptive fields and encode the unisensory likelihood functions. Hence, after training, each unisensory area realizes a maximum likelihood estimate of stimulus position (auditory or visual). In cross-modal conditions, the same learning rule can encode information on prior probability into the cross-modal synapses. Computer simulations confirm the theoretical results and show that the proposed network can realize a maximum likelihood estimate of auditory (or visual) positions in unimodal conditions and a Bayesian estimate, with moderate deviations from optimality, in cross-modal conditions. Furthermore, the model explains the ventriloquism illusion and, looking at the activity in the multimodal neurons, explains the automatic reweighting of auditory and visual inputs on a trial-by-trial basis, according to the reliability of the individual cues.

Author(s):  
Russell Cheng

This chapter examines the well-known Box-Cox method, which transforms a sample of non-normal observations into approximately normal form. Two non-standard aspects are highlighted. First, the likelihood of the transformed sample has an unbounded maximum, so that the maximum likelihood estimate is not consistent. The usually suggested remedy is to assume grouped data so that the sample becomes multinomial. An alternative method is described that uses a modified likelihood similar to the spacings function. This eliminates the infinite likelihood problem. The second problem is that the power transform used in the Box-Cox method is left-bounded so that the transformed observations cannot be exactly normal. This biases estimates of observational probabilities in an uncertain way. Moreover, the distributions fitted to the observations are not necessarily unimodal. A simple remedy is to assume the transformed observations have a left-bounded distribution, like the exponential; this is discussed in detail, and a numerical example given.


Sign in / Sign up

Export Citation Format

Share Document