scholarly journals Virtual-Audio Aided Visual Search on a Desktop Display

Author(s):  
Clayton Rothwell ◽  
Griffin Romigh ◽  
Brian Simpson

As visual display complexity grows, visual cues and alerts may become less salient and therefore less effective. Although the auditory system’s resolution is rather coarse relative to the visual system, there is some evidence for virtual spatialized audio to benefit visual search on a small frontal region, such as a desktop monitor. Two experiments examined if search times could be reduced compared to visual-only search through spatial auditory cues rendered using one of two methods: individualized or generic head-related transferfunctions. Results showed the cue type interacted with display complexity, with larger reductions compared to visual-only search as set size increased. For larger set sizes, individualized cues were significantly better than generic cues overall. Across all set sizes, individualized cues were better than generic cues for cueing eccentric elevations (>±8°). Where performance must be maximized, designers should use individualized virtual audio if at all possible, even in small frontal region within the field of view.

Author(s):  
Bartholomew Elias

Since the auditory system is not spatially restricted like the visual system, spatial auditory cues can provide information regarding object position, velocity, and trajectory beyond the field of view. A laboratory experiment was conducted to demonstrate that visual displays can be augmented with dynamic spatial auditory cues that provide information regarding the motion characteristics of unseen objects. In this study, dynamic spatial auditory cues presented through headphones conveyed preview information regarding target position, velocity, and trajectory beyond the field of view in a dynamic visual search task in which subjects acquired and identified moving visual targets that traversed a display cluttered with varying numbers of moving distractors. The provision of spatial auditory preview significantly reduced response times to acquire and identify the visual targets and significantly reduced error rates, especially in cases when the visual display load was high. These findings demonstrate that providing dynamic spatial auditory preview cues is a viable mechanism for augmenting visual search performance in dynamic task environments.


Author(s):  
Rachel J. Cunio ◽  
David Dommett ◽  
Joseph Houpt

Maintaining spatial awareness is a primary concern for operators, but relying only on visual displays can cause visual system overload and lead to performance decrements. Our study examined the benefits of providing spatialized auditory cues for maintaining visual awareness as a method of combating visual system overload. We examined visual search performance of seven participants in an immersive, dynamic (moving), three-dimensional, virtual reality environment both with no cues, non-masked, spatialized auditory cues, and masked, spatialized auditory cues. Results indicated a significant reduction in visual search time from the no-cue condition when either auditory cue type was presented, with the masked auditory condition slower. The results of this study can inform attempts to improve visual search performance in operational environments, such as determining appropriate display types for providing spatial information.


1979 ◽  
Vol 23 (1) ◽  
pp. 301-304 ◽  
Author(s):  
Gavan Lintern

An aircraft simulator with a closed-loop computer-generated visual display, was used to teach flight-naive subjects to land. A control training condition in which subjects learned to land with reference to a skeletal airport scene consisting of a horizon, runway, centerline, and aiming bar, was tested against training with constantly augmented feedback, adaptively augmented feedback, and a flightpath tracking display. A simulator-to-simulator transfer-of-training design showed that adaptively trained subjects performed best in a transfer task that was identical to the control group's training condition. Several subjects attempted six landings in a light airplane after they had completed their experimental work in the simulator. They performed better than another group of subjects that had not had any landing practice in the simulator.


1996 ◽  
Vol 5 (3) ◽  
pp. 330-345 ◽  
Author(s):  
Edward J. Rinalducci

This paper provides an overview of the literature on the visual system, placing special emphasis on those visual characteristics regarded as necessary to produce adequate visual fidelity in virtual environments. These visual cues apply to the creation of various virtual environments including those involving flying, driving, sailing, or walking. A variety of cues are examined, in particular, motion, color, stereopsis, pictorial and secondary cues, physiological cues, texture, vertical development, luminance, field-of-view, and spatial resolution. Conclusions and recommendations for research are also presented.


Author(s):  
Amit Barde Brock ◽  
Matt Ward ◽  
William S. Helton ◽  
Mark Billinghurst

Attention redirection trials were carried out using a wearable interface incorporating auditory and visual cues. Visual cues were delivered via the screen on the Recon Jet – a wearable computer resembling a pair of glasses – while auditory cues were delivered over a bone conduction headset. Cueing conditions included the delivery of individual cues, both auditory and visual, and in combination with each other. Results indicate that the use of an auditory cue drastically decreases target acquisition times. This is true especially for targets that fall outside the visual field of view. While auditory cues showed no difference when paired with any of the visual cueing conditions for targets within the field of view of the user, for those outside the field of view a significant improvement in performance was observed. The static visual cue paired with the binaurally spatialised, dynamic auditory cue appeared to provide the best performance in comparison to any other cueing conditions. In the absence of a visual cue, the binaurally spatialised, dynamic auditory cue performed the best.


Author(s):  
Kit W. Cho

Abstract. Words rated for their survival relevance are remembered better than when rated using other well-known memory mnemonics. This finding, which is known as the survival advantage effect and has been replicated in many studies, suggests that our memory systems are molded by natural selection pressures. In two experiments, the present study used a visual search task to examine whether there is likewise a survival advantage for our visual systems. Participants rated words for their survival relevance or for their pleasantness before locating that object’s picture in a search array with 8 or 16 objects. Although there was no difference in search times among the two rating scenarios when set size was 8, survival processing reduced visual search times when set size was 16. These findings reflect a search efficiency effect and suggest that similar to our memory systems, our visual systems are also tuned toward self-preservation.


1977 ◽  
Vol 44 (2) ◽  
pp. 647-655 ◽  
Author(s):  
Kathryn M. Yorkston ◽  
Robert C. Marshall ◽  
Myrtice R. Butler

Two groups of aphasics were administered an auditory comprehension task under conditions of 0-, 5-, and 10-sec. imposed delay of response. The auditory-visual group received auditory and visual cues; the auditory group received only auditory cues. Comprehension for the auditory-visual group was significantly better than for the auditory group. Increase in delay time significantly improved comprehension for the auditory-visual group but not for the auditory group.


2018 ◽  
Author(s):  
Bernt Skottun

In tests of visual search it is found that the response time increases with set size under some stimulus conditions whereas it may be relatively independent of set size under other conditions. The present study investigates interference in the kinds of stimuli employed in such experiments. Interference takes place in the stimuli and does not depend on the visual system or on attention. Two scenarios were explored: (1) Interference between all elements, and (2) interference between a target and a single background made up of a set of elements. In the former case interference increases monotonically with set size. In the latter case there is a mild amount of interference which remains relatively independent of set size. By assuming the response time increases with the inverse of stimulus power (response time ∝ 1/(1 - interference)) it is found that when there is interference between all the elements the simulated response time increases monotonically with set size. On the other hand, when the interference is between a target and a single background the simulated response time remains largely unaffected by set size.


Author(s):  
Adam F. Werner ◽  
Jamie C. Gorman

Objective This study examines visual, auditory, and the combination of both (bimodal) coupling modes in the performance of a two-person perceptual-motor task, in which one person provides the perceptual inputs and the other the motor inputs. Background Parking a plane or landing a helicopter on a mountain top requires one person to provide motor inputs while another person provides perceptual inputs. Perceptual inputs are communicated either visually, auditorily, or through both cues. Methods One participant drove a remote-controlled car around an obstacle and through a target, while another participant provided auditory, visual, or bimodal cues for steering and acceleration. Difficulty was manipulated using target size. Performance (trial time, path variability), cue rate, and spatial ability were measured. Results Visual coupling outperformed auditory coupling. Bimodal performance was best in the most difficult task condition but also high in the easiest condition. Cue rate predicted performance in all coupling modes. Drivers with lower spatial ability required a faster auditory cue rate, whereas drivers with higher ability performed best with a lower rate. Conclusion Visual cues result in better performance when only one coupling mode is available. As predicted by multiple resource theory, when both cues are available, performance depends more on auditory cueing. In particular, drivers must be able to transform auditory cues into spatial actions. Application Spotters should be trained to provide an appropriate cue rate to match the spatial ability of the driver or pilot. Auditory cues can enhance visual communication when the interpersonal task is visual with spatial outputs.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Martin Saveski ◽  
Edmond Awad ◽  
Iyad Rahwan ◽  
Manuel Cebrian

AbstractAs groups are increasingly taking over individual experts in many tasks, it is ever more important to understand the determinants of group success. In this paper, we study the patterns of group success in Escape The Room, a physical adventure game in which a group is tasked with escaping a maze by collectively solving a series of puzzles. We investigate (1) the characteristics of successful groups, and (2) how accurately humans and machines can spot them from a group photo. The relationship between these two questions is based on the hypothesis that the characteristics of successful groups are encoded by features that can be spotted in their photo. We analyze >43K group photos (one photo per group) taken after groups have completed the game—from which all explicit performance-signaling information has been removed. First, we find that groups that are larger, older and more gender but less age diverse are significantly more likely to escape. Second, we compare humans and off-the-shelf machine learning algorithms at predicting whether a group escaped or not based on the completion photo. We find that individual guesses by humans achieve 58.3% accuracy, better than random, but worse than machines which display 71.6% accuracy. When humans are trained to guess by observing only four labeled photos, their accuracy increases to 64%. However, training humans on more labeled examples (eight or twelve) leads to a slight, but statistically insignificant improvement in accuracy (67.4%). Humans in the best training condition perform on par with two, but worse than three out of the five machine learning algorithms we evaluated. Our work illustrates the potentials and the limitations of machine learning systems in evaluating group performance and identifying success factors based on sparse visual cues.


Sign in / Sign up

Export Citation Format

Share Document