scholarly journals Visual Information Requirements for Remotely Supervised Autonomous Agricultural Machines

2020 ◽  
Vol 10 (8) ◽  
pp. 2794 ◽  
Author(s):  
Uduak Edet ◽  
Daniel Mann

A study to determine the visual requirements for a remote supervisor of an autonomous sprayer was conducted. Observation of a sprayer operator identified 9 distinct “look zones” that occupied his visual attention, with 39% of his time spent viewing the look zone ahead of the sprayer. While observation of the sprayer operator was being completed, additional GoPro cameras were used to record video of the sprayer in operation from 10 distinct perspectives (some look zones were visible from the operator’s seat, but other look zones were selected to display other regions of the sprayer that might be of interest to a sprayer operator). In a subsequent laboratory study, 29 experienced sprayer operators were recruited to view and comment on video clips selected from the video footage collected during the initial ride-along. Only the two views from the perspective of the operator’s seat were rated highly as providing important information even though participants were able to identify relevant information from all ten of the video clips. Generally, participants used the video clips to obtain information about the boom status, the location and movement of the sprayer within the field, the weather conditions (especially the wind), obstacles to be avoided, crop conditions, and field conditions. Sprayer operators with more than 15 years of experience provided more insightful descriptions of the video clips than their less experienced peers. Designers can influence which features the user will perceive by positioning the camera such that those specific features are prominent in the camera’s field of view. Overall, experienced sprayer operators preferred the concept of presenting visual information on an automation interface using live video rather than presenting that same information using some type of graphical display using icons or symbols.

2021 ◽  
Vol 11 (4) ◽  
pp. 3023-3029
Author(s):  
Muhammad Junaid ◽  
Luqman Shah ◽  
Ali Imran Jehangiri ◽  
Fahad Ali Khan ◽  
Yousaf Saeed ◽  
...  

With each passing day resolutions of still image/video cameras are on the rise. This amelioration in resolutions has the potential to extract useful information on the view opposite the photographed subjects from their reflecting parts. Especially important is the idea to capture images formed on the eyes of photographed people and animals. The motivation behind this research is to explore the forensic importance of the images/videos to especially analyze the reflections of the background of the camera. This analysis may include extraction/ detection/recognition of the objects in front of the subjects but on the back of the camera. In the national context such videos/photographs are not rare and, specifically speaking, an abductee’s video footage at a good resolution may give some important clues to the identity of the person who kidnapped him/her. Our aim would be to extract visual information formed in human eyes from still images as well as from video clips. After extraction, our next task would be to recognize the extracted visual information. Initially our experiments would be limited on characters’ extraction and recognition, including characters of different styles and font sizes (computerized) as well as hand written. Although varieties of Optical Character Recognition (OCR) tools are available for characters’ extraction and recognition but, the problem is that they only provide results for clear images (zoomed).


2020 ◽  
Vol 228 (1) ◽  
pp. 43-49 ◽  
Author(s):  
Michael Kossmeier ◽  
Ulrich S. Tran ◽  
Martin Voracek

Abstract. Currently, dedicated graphical displays to depict study-level statistical power in the context of meta-analysis are unavailable. Here, we introduce the sunset (power-enhanced) funnel plot to visualize this relevant information for assessing the credibility, or evidential value, of a set of studies. The sunset funnel plot highlights the statistical power of primary studies to detect an underlying true effect of interest in the well-known funnel display with color-coded power regions and a second power axis. This graphical display allows meta-analysts to incorporate power considerations into classic funnel plot assessments of small-study effects. Nominally significant, but low-powered, studies might be seen as less credible and as more likely being affected by selective reporting. We exemplify the application of the sunset funnel plot with two published meta-analyses from medicine and psychology. Software to create this variation of the funnel plot is provided via a tailored R function. In conclusion, the sunset (power-enhanced) funnel plot is a novel and useful graphical display to critically examine and to present study-level power in the context of meta-analysis.


Perception ◽  
2017 ◽  
Vol 46 (12) ◽  
pp. 1412-1426 ◽  
Author(s):  
Elmeri Syrjänen ◽  
Marco Tullio Liuzza ◽  
Håkan Fischer ◽  
Jonas K. Olofsson

Disgust is a core emotion evolved to detect and avoid the ingestion of poisonous food as well as the contact with pathogens and other harmful agents. Previous research has shown that multisensory presentation of olfactory and visual information may strengthen the processing of disgust-relevant information. However, it is not known whether these findings extend to dynamic facial stimuli that changes from neutral to emotionally expressive, or if individual differences in trait body odor disgust may influence the processing of disgust-related information. In this preregistered study, we tested whether a classification of dynamic facial expressions as happy or disgusted, and an emotional evaluation of these facial expressions, would be affected by individual differences in body odor disgust sensitivity, and by exposure to a sweat-like, negatively valenced odor (valeric acid), as compared with a soap-like, positively valenced odor (lilac essence) or a no-odor control. Using Bayesian hypothesis testing, we found evidence that odors do not affect recognition of emotion in dynamic faces even when body odor disgust sensitivity was used as moderator. However, an exploratory analysis suggested that an unpleasant odor context may cause faster RTs for faces, independent of their emotional expression. Our results further our understanding of the scope and limits of odor effects on facial perception affect and suggest further studies should focus on reproducibility, specifying experimental circumstances where odor effects on facial expressions may be present versus absent.


2019 ◽  
Author(s):  
Roy S Hessels

Gaze – where one looks, how long, and when – plays an essential part in human social behaviour. While many aspects of social gaze have been reviewed, there is no comprehensive review or theoretical framework that describes how gaze to faces supports face-to-face interaction. In this review, I address the following questions: (1) When does gaze need to be allocated to a particular region of a face in order to provide the relevant information for successful interaction; (2) How do humans look at other people, and faces in particular, regardless of whether gaze needs to be directed at a particular region to acquire the relevant visual information; (3) How does gaze support the regulation of interaction? The work reviewed spans psychophysical research, observational research and eye-tracking research in both lab-based and interactive contexts. Based on the literature overview, I sketch a framework for future research based on dynamic systems theory. The framework holds that gaze should be investigated in relation to sub-states of the interaction, encompassing sub-states of the interactors, the content of the interaction as well as the interactive context. The relevant sub-states for understanding gaze in interaction vary over different timescales from microgenesis to ontogenesis and phylogenesis. The framework has important implications for vision science, psychopathology, developmental science and social robotics.


2014 ◽  
Vol 2014 ◽  
pp. 1-23 ◽  
Author(s):  
Francisco Amorós ◽  
Luis Payá ◽  
Oscar Reinoso ◽  
Walterio Mayol-Cuevas ◽  
Andrew Calway

In this work we present a topological map building and localization system for mobile robots based on global appearance of visual information. We include a comparison and analysis of global-appearance techniques applied to wide-angle scenes in retrieval tasks. Next, we define multiscale analysis, which permits improving the association between images and extracting topological distances. Then, a topological map-building algorithm is proposed. At first, the algorithm has information only of some isolated positions of the navigation area in the form of nodes. Each node is composed of a collection of images that covers the complete field of view from a certain position. The algorithm solves the node retrieval and estimates their spatial arrangement. With these aims, it uses the visual information captured along some routes that cover the navigation area. As a result, the algorithm builds a graph that reflects the distribution and adjacency relations between nodes (map). After the map building, we also propose a route path estimation system. This algorithm takes advantage of the multiscale analysis. The accuracy in the pose estimation is not reduced to the nodes locations but also to intermediate positions between them. The algorithms have been tested using two different databases captured in real indoor environments under dynamic conditions.


2020 ◽  
Vol 12 (13) ◽  
pp. 5308 ◽  
Author(s):  
Jiaming Shi ◽  
Changxu Wu ◽  
Xiuying Qian

To make safe road-crossing decisions, it is necessary for pedestrians to accurately estimate the speed and stopping distance of approaching vehicles. Accordingly, the objective of our study was to examine the effects of multiple factors, such as weather conditions, context time (day or night), and illuminance of the roads, on older pedestrians’ (>60 years old) speed perception and stopping distance estimation of approaching vehicles. The participants in this study included 48 older participants who were asked to estimate the speed and stopping distance of approaching vehicles based on 12 s video clips that were selected from natural conditions. The results revealed that actual speeds, weather, context time, and lighting conditions played important roles in the performance of the participants. Compared with young adults, older pedestrians were found to have smaller accurate estimation intervals that varied by multidimensional influencing factors and thus resulted in missing road-crossing opportunities at lower vehicles’ speeds and increasing road-crossing dangers at higher speeds. The older pedestrians’ performance with respect to speed perception and stopping distance estimation is modeled using a regression model with a complex level of tasks. These models can be used by engineers when establishing speed limits and lighting conditions in the areas with senior residents.


2019 ◽  
Vol 4 (1) ◽  
pp. 116-137
Author(s):  
Theresa Giorza

A public park adjacent to an inner-city preschool invites children and their teacher into new encounters with the world, literacy and themselves. The park and preschool are situated in the inner-city of Johannesburg, South Africa. In this article, the researcher performs as mutated-modest-witness of events that unfold in lively materialdiscursive encounters between children, grass, friendship, a pen, cement table, sand, sticks, the alphabet and daylight. The agential realism of Karen Barad and the nomadic thinking of Deleuze and Guattari offer ways of re-imagining ‘the child in the park’. Diffracting with repeated viewings of video clips the researcher finds that forward and reverse movement and stops in different moments throughout repeated viewings of the same video footage produces different and new ‘stories’ about the events and the children involved. Conceptions of ‘child’ as literacy learner and of researcher-as-writer mutate through this diffraction which instantiates a non-representational videography practice.


2010 ◽  
Vol 103 (2) ◽  
pp. 1057-1065 ◽  
Author(s):  
Edmund Wascher ◽  
Christian Beste

The ability to notice relevant visual information has been assumed to be determined both by the relative salience of relevant information compared with distracters within a given display and by voluntary allocation of attention toward intended goals. A dominance of either of these two mechanisms in stimulus processing has been claimed by different theories. A central question in this context is to what degree and how task irrelevant signals can influence processing of target information. In the present study, participants had to detect a luminance change in various conditions among others against an irrelevant orientation change. The saliency of the latter was systematically varied and was found to be predictive for the proportion of detected information when relevant and irrelevant information were spatially separated but not when they overlapped. Weighting and competition of incoming signals was reflected in the amplitude of the N1pc component of the event-related potential. Initial orientation of attention toward the irrelevant element had to be followed by a reallocation process, reflected in an N2pc. The control of conflicting information additionally evoked a fronto-central N2 that varied with the amount of competition induced. Thus the data support models that assume that attention is a dynamic interplay of bottom-up and top-down processes that may be mediated via a common dynamic neural network.


2016 ◽  
Vol 2016 ◽  
pp. 1-21 ◽  
Author(s):  
L. Payá ◽  
O. Reinoso ◽  
Y. Berenguer ◽  
D. Úbeda

Nowadays, the design of fully autonomous mobile robots is a key discipline. Building a robust model of the unknown environment is an important ability the robot must develop. Using this model, this robot must be able to estimate its current position and to navigate to the target points. The use of omnidirectional vision sensors is usual to solve these tasks. When using this source of information, the robot must extract relevant information from the scenes both to build the model and to estimate its position. The possible frameworks include the classical approach of extracting and describing local features or working with the global appearance of the scenes, which has emerged as a conceptually simple and robust solution. While feature-based techniques have been extensively studied in the literature, appearance-based ones require a full comparative evaluation to reveal the performance of the existing methods and to tune correctly their parameters. This work carries out a comparative evaluation of four global-appearance techniques in map building tasks, using omnidirectional visual information as the only source of data from the environment.


Sign in / Sign up

Export Citation Format

Share Document