Effects of Stimulus Eccentricity on the Perception of Visually Induced Self-motion Facilitated by Simulated Viewpoint Jitter

2012 ◽  
Vol 25 (6) ◽  
pp. 647-654 ◽  
Author(s):  
Shinji Nakamura

The present investigation aimed to examine the effects of stimulus eccentricity in the facilitation of vection by a jittering visual inducer. A psychophysical experiment revealed that the central region of the visual field is more critical in facilitation by perspective viewpoint jitter than the peripheral area. The results suggest that the perceptual mechanism underlying the facilitation by jitter may be different from that responsible for generating standard vection from non-jittering visual motion, because the effects of stimulus eccentricity were quite different in these two situations.

2013 ◽  
Vol 26 (3) ◽  
pp. 277-285 ◽  
Author(s):  
Shinji Nakamura

It has been repeatedly reported that visual stimuli containing a jittering/oscillating motion component can induce self-motion perception more strongly than a pure radial expansion pattern. A psychophysical experiment with 11 observers revealed that the additional accelerating components of the visual motion have to be convoluted with the motion of the main-axis to facilitate self-motion perception; additional motion presented in an isolated fashion impairs the perception of self-motion. These results are inconsistent with a simple hypothesis about the perceptual mechanism underlying the advantage of jitter/oscillation, which assumes that the accelerating component induces an additional self-motion independently of the main motion at the first stage, and then the two self-motions induced by the main motion and the additional component become integrated.


2020 ◽  
Vol 1 ◽  
Author(s):  
Kanon Fujimoto ◽  
Hiroshi Ashida

Humans perceive self-motion using multisensory information, while vision has a dominant role as is utilized in virtual reality (VR) technologies. Previous studies reported that visual motion presented in the lower visual field (LoVF) induces stronger illusion of self-motion (vection) as compared with the upper visual field (UVF). However, it was still unknown whether the LoVF superiority in vection was based on the retinotopic frame, or rather related to the environmental frame of reference. Here, we investigated the influences of retinotopic and environmental frames on the LoVF superiority of vection. We presented a planer surface along the depth axis in one of four visual fields (upper, lower, right, or left). The texture on the surface moved forward or backward. Participants reported vection while observing the visual stimulus through a VR head mounted display (HMD) in the sitting posture or lateral recumbent position. Results showed that the visual motion induced stronger vection when presented in the LoVF compared with the UVF in both postures. Notably, the vection rating in LoVF was stronger in the sitting than in the recumbent. Moreover, recumbent participants reported stronger vection when the stimulus was presented in the gravitationally lower field than in the gravitationally upper field. These results demonstrate contribution of multiple spatial frames on self-motion perception and imply the importance of ground surface.


2004 ◽  
Vol 14 (5) ◽  
pp. 375-385 ◽  
Author(s):  
E.L. Groen ◽  
W. Bles

We examined to what extent body tilt may augment the perception of visually simulated linear self acceleration. Fourteen subjects judged visual motion profiles of fore-aft motion at four different frequencies between 0.04âĂŞ0.33 Hz, and at three different acceleration amplitudes (0.44, 0.88 and 1.76 m / s 2 ). Simultaneously, subjects were tilted backward and forward about their pitch axis. The amplitude of pitch tilt was systematically varied. Using a two-alternative-forced-choice paradigm, psychometric curves were calculated in order to determine: 1) the minimum tilt amplitude required to generate a linear self-motion percept in more than 50% of the cases, and 2) the maximum tilt amplitude at which rotation remains sub-threshold in more than 50% of the cases. The results showed that the simulation of linear self motion became more realistic with the application of whole body tilt, as long as the tilt rate remained under the detection threshold of about 3 deg/s. This value is in close agreement with the empirical rate limit commonly used in flight simulation. The minimum required motion cue was inversely proportional to stimulus frequency, and increased with the amplitude of the visual displacement (rather than acceleration). As a consequence, the range of useful tilt stimuli became more critical with increasing stimulus frequency. We conclude that this psychophysical approach reveals valid parameters for motion driving algorithms used in motion base simulators.


2006 ◽  
Vol 95 (6) ◽  
pp. 3712-3726 ◽  
Author(s):  
Frédéric V. Barthélemy ◽  
Ivo Vanzetta ◽  
Guillaume S. Masson

Visual neurons integrate information over a finite part of the visual field with high selectivity. This classical receptive field is modulated by peripheral inputs that play a role in both neuronal response normalization and contextual modulations. However, the consequences of these properties for visuomotor transformations are yet incompletely understood. To explore those, we recorded short-latency ocular following responses in humans to large center-only and center-surround stimuli. We found that eye movements are triggered by a mechanism that integrates motion over a restricted portion of the visual field, the size of which depends on stimulus contrast and increases as a function of time after response onset. We also found evidence for a strong nonisodirectional center-surround organization, responsible for normalizing the central, driving input so that motor responses are set to their most linear contrast dynamics. Such response normalization is delayed about 20 ms relative to tracking onset, gradually builds up over time, and is partly tuned for surround orientation/direction. These results outline the spatiotemporal organization of a behavioral receptive field, which might reflect a linear integration among subpopulations of cortical visual motion detectors.


1988 ◽  
Vol 60 (3) ◽  
pp. 940-965 ◽  
Author(s):  
M. R. Dursteler ◽  
R. H. Wurtz

1. Previous experiments have shown that punctate chemical lesions within the middle temporal area (MT) of the superior temporal sulcus (STS) produce deficits in the initiation and maintenance of pursuit eye movements (10, 34). The present experiments were designed to test the effect of such chemical lesions in an area within the STS to which MT projects, the medial superior temporal area (MST). 2. We injected ibotenic acid into localized regions of MST, and we observed two deficits in pursuit eye movements, a retinotopic deficit and a directional deficit. 3. The retinotopic deficit in pursuit initiation was characterized by the monkey's inability to match eye speed to target speed or to adjust the amplitude of the saccade made to acquire the target to compensate for target motion. This deficit was related to the initiation of pursuit to targets moving in any direction in the visual field contralateral to the side of the brain with the lesion. This deficit was similar to the deficit we found following damage to extrafoveal MT except that the affected area of the visual field frequently extended throughout the entire contralateral visual field tested. 4. The directional deficit in pursuit maintenance was characterized by a failure to match eye speed to target speed once the fovea had been brought near the moving target. This deficit occurred only when the target was moving toward the side of the lesion, regardless of whether the target began to move in the ipsilateral or contralateral visual field. There was no deficit in the amplitude of saccades made to acquire the target, or in the amplitude of the catch-up saccades made to compensate for the slowed pursuit. The directional deficit is similar to the one we described previously following chemical lesions of the foveal representation in the STS. 5. Retinotopic deficits resulted from any of our injections in MST. Directional deficits resulted from lesions limited to subregions within MST, particularly lesions that invaded the floor of the STS and the posterior bank of the STS just lateral to MT. Extensive damage to the densely myelinated area of the anterior bank or to the posterior parietal area on the dorsal lip of the anterior bank produced minimal directional deficits. 6. We conclude that damage to visual motion processing in MST underlies the retinotopic pursuit deficit just as it does in MT. MST appears to be a sequential step in visual motion processing that occurs before all of the visual motion information is transmitted to the brainstem areas related to pursuit.(ABSTRACT TRUNCATED AT 400 WORDS)


2020 ◽  
Author(s):  
Maria-Bianca Leonte ◽  
Aljoscha Leonhardt ◽  
Alexander Borst ◽  
Alex S. Mauss

AbstractVisual motion detection is among the best understood neuronal computations. One assumed behavioural role is to detect self-motion and to counteract involuntary course deviations, extensively investigated in tethered walking or flying flies. In free flight, however, any deviation from a straight course is signalled by both the visual system as well as by proprioceptive mechanoreceptors called ‘halteres’, which are the equivalent of the vestibular system in vertebrates. Therefore, it is yet unclear to what extent motion vision contributes to course control, or whether straight flight is completely controlled by proprioceptive feedback from the halteres. To answer these questions, we genetically rendered flies motion-blind by blocking their primary motion-sensitive neurons and quantified their free-flight performance. We found that such flies have difficulties maintaining a straight flight trajectory, much like control flies in the dark. By unilateral wing clipping, we generated an asymmetry in propulsory force and tested the ability of flies to compensate for this perturbation. While wild-type flies showed a remarkable level of compensation, motion-blind animals exhibited pronounced circling behaviour. Our results therefore unequivocally demonstrate that motion vision is necessary to fly straight under realistic conditions.


i-Perception ◽  
2021 ◽  
Vol 12 (6) ◽  
pp. 204166952110557
Author(s):  
Diederick C. Niehorster

The concept of optic flow, a global pattern of visual motion that is both caused by and signals self-motion, is canonically ascribed to James Gibson's 1950 book “ The Perception of the Visual World.” There have, however, been several other developments of this concept, chiefly by Gwilym Grindley and Edward Calvert. Based on rarely referenced scientific literature and archival research, this article describes the development of the concept of optic flow by the aforementioned authors and several others. The article furthermore presents the available evidence for interactions between these authors, focusing on whether parts of Gibson's proposal were derived from the work of Grindley or Calvert. While Grindley's work may have made Gibson aware of the geometrical facts of optic flow, Gibson's work is not derivative of Grindley's. It is furthermore shown that Gibson only learned of Calvert's work in 1956, almost a decade after Gibson first published his proposal. In conclusion, the development of the concept of optic flow presents an intriguing example of convergent thought in the progress of science.


2021 ◽  
Vol 118 (32) ◽  
pp. e2106235118
Author(s):  
Reuben Rideaux ◽  
Katherine R. Storrs ◽  
Guido Maiello ◽  
Andrew E. Welchman

Sitting in a static railway carriage can produce illusory self-motion if the train on an adjoining track moves off. While our visual system registers motion, vestibular signals indicate that we are stationary. The brain is faced with a difficult challenge: is there a single cause of sensations (I am moving) or two causes (I am static, another train is moving)? If a single cause, integrating signals produces a more precise estimate of self-motion, but if not, one cue should be ignored. In many cases, this process of causal inference works without error, but how does the brain achieve it? Electrophysiological recordings show that the macaque medial superior temporal area contains many neurons that encode combinations of vestibular and visual motion cues. Some respond best to vestibular and visual motion in the same direction (“congruent” neurons), while others prefer opposing directions (“opposite” neurons). Congruent neurons could underlie cue integration, but the function of opposite neurons remains a puzzle. Here, we seek to explain this computational arrangement by training a neural network model to solve causal inference for motion estimation. Like biological systems, the model develops congruent and opposite units and recapitulates known behavioral and neurophysiological observations. We show that all units (both congruent and opposite) contribute to motion estimation. Importantly, however, it is the balance between their activity that distinguishes whether visual and vestibular cues should be integrated or separated. This explains the computational purpose of puzzling neural representations and shows how a relatively simple feedforward network can solve causal inference.


PLoS ONE ◽  
2021 ◽  
Vol 16 (12) ◽  
pp. e0261266
Author(s):  
Maëlle Tixier ◽  
Stéphane Rousset ◽  
Pierre-Alain Barraud ◽  
Corinne Cian

A large body of research has shown that visually induced self-motion (vection) and cognitive processing may interfere with each other. The aim of this study was to assess the interactive effects of a visual motion inducing vection (uniform motion in roll) versus a visual motion without vection (non-uniform motion) and long-term memory processing using the characteristics of standing posture (quiet stance). As the level of interference may be related to the nature of the cognitive tasks used, we examined the effect of visual motion on a memory task which requires a spatial process (episodic recollection) versus a memory task which does not require this process (semantic comparisons). Results confirm data of the literature showing that compensatory postural response in the same direction as background motion. Repeatedly watching visual uniform motion or increasing the cognitive load with a memory task did not decrease postural deviations. Finally, participants were differentially controlling their balance according to the memory task but this difference was significant only in the vection condition and in the plane of background motion. Increased sway regularity (decreased entropy) combined with decreased postural stability (increase variance) during vection for the episodic task would indicate an ineffective postural control. The different interference of episodic and semantic memory on posture during visual motion is consistent with the involvement of spatial processes during episodic memory recollection. It can be suggested that spatial disorientation due to visual roll motion preferentially interferes with spatial cognitive tasks, as spatial tasks can draw on resources expended to control posture.


2017 ◽  
Vol 30 (1) ◽  
pp. 65-90 ◽  
Author(s):  
Séamas Weech ◽  
Nikolaus F. Troje

Studies of the illusory sense of self-motion elicited by a moving visual surround (‘vection’) have revealed key insights about how sensory information is integrated. Vection usually occurs after a delay of several seconds following visual motion onset, whereas self-motion in the natural environment is perceived immediately. It has been suggested that this latency relates to the sensory mismatch between visual and vestibular signals at motion onset. Here, we tested three techniques with the potential to reduce sensory mismatch in order to shorten vection onset latency: noisy galvanic vestibular stimulation (GVS) and bone conducted vibration (BCV) at the mastoid processes, and body vibration applied to the lower back. In Experiment 1, we examined vection latency for wide field visual rotations about the roll axis and applied a burst of stimulation at the start of visual motion. Both GVS and BCV reduced vection latency by two seconds compared to the control condition, whereas body vibration had no effect on latency. In Experiment 2, the visual stimulus rotated about the pitch, roll, or yaw axis and we found a similar facilitation of vection by both BCV and GVS in each case. In a control experiment, we confirmed that air-conducted sound administered through headphones was not sufficient to reduce vection onset latency. Together the results suggest that noisy vestibular stimulation facilitates vection, likely due to an upweighting of visual information caused by a reduction in vestibular sensory reliability.


Sign in / Sign up

Export Citation Format

Share Document