Visual Perception for a Partner Robot Based on Computational Intelligence

Author(s):  
Indra Adji Sulistijono ◽  
◽  
Naoyuki Kubota ◽  
◽  
◽  
...  

We propose computational intelligence for partner robot perception in which the robot requires the capability of visual perception to interact with human beings. Basically, robots should conduct moving object extraction, clustering, and classification for visual perception used in interactions with human beings. We propose total human visual tracking by long-term memory, k-means, self-organizing map, and a fuzzy controller is used for movement output. Experimental results show that the partner robot can conduct the human visual tracking.

Author(s):  
Kai Essig ◽  
Oleg Strogan ◽  
Helge Ritter ◽  
Thomas Schack

Various computational models of visual attention rely on the extraction of salient points or proto-objects, i.e., discrete units of attention, computed from bottom-up image features. In recent years, different solutions integrating top-down mechanisms were implemented, as research has shown that although eye movements initially are solely influenced by bottom-up information, after some time goal driven (high-level) processes dominate the guidance of visual attention towards regions of interest (Hwang, Higgins & Pomplun, 2009). However, even these improved modeling approaches are unlikely to generalize to a broader range of application contexts, because basic principles of visual attention, such as cognitive control, learning and expertise, have thus far not sufficiently been taken into account (Tatler, Hayhoe, Land & Ballard, 2011). In some recent work, the authors showed the functional role and representational nature of long-term memory structures for human perceptual skills and motor control. Based on these findings, the chapter extends a widely applied saliency-based model of visual attention (Walther & Koch, 2006) in two ways: first, it computes the saliency map using the cognitive visual attention approach (CVA) that shows a correspondence between regions of high saliency values and regions of visual interest indicated by participants’ eye movements (Oyekoya & Stentiford, 2004). Second, it adds an expertise-based component (Schack, 2012) to represent the influence of the quality of mental representation structures in long-term memory (LTM) and the roles of learning on the visual perception of objects, events, and motor actions.


2017 ◽  
Vol 14 (02) ◽  
pp. 1702001 ◽  
Author(s):  
Young-Jae Ryoo ◽  
Takahiro Yamanoi

The special issue topics focus on the computational intelligence and its application for robotics. Its areas reach out comprehensive ranges; context-awareness software, omnidirectional walking and fuzzy controller of dynamic walking for humanoid robots, pet robots for treatment of ASD children, fuzzy logic control, enhanced simultaneous localization and mapping, fuzzy line tracking for mobile robots, and so on. Computational intelligence (CI) is a method of performing like humans. Generally computational intelligence means the ability of a computer to learn a specific task from data or experimental results. Meanwhile robotic system has many limits to behave like human beings. The robotic system might be too complex for mathematical reasoning, it might contain some uncertainties during the process, or the process might simply be stochastic in real life. Real-life problems cannot be translated into binary code for computers to process it. Computational intelligence might solve such problems.


2017 ◽  
Author(s):  
Jason Samaha ◽  
Bradley R. Postle

AbstractAdaptive behavior depends on the ability to accurately introspect about one’s own performance. Whether this metacognitive ability is supported by the same mechanisms across different tasks has thus far been investigated with a focus on correlating metacognitive accuracy between perception and long-term memory paradigms. Here, we investigated the relationship between metacognition of visual perception and metacognition of visual short-term memory (VSTM), a cognitive function thought to be more intimately related to visual processing. Experiments 1 and 2 required subjects to estimate the perceived or remembered orientation of a grating stimulus and rate their confidence. We observed strong positive correlations between individual differences in metacognitive accuracy between the two tasks. This relationship was not accounted for by individual differences in task performance or average confidence, and was present across two different metrics of metacognition and in both experiments. A model-based analysis of data from a third experiment showed that a cross-domain correlation only emerged when both tasks shared the same task-relevant stimulus feature. That is, metacognition for perception and VSTM were correlated when both tasks required orientation judgments, but not when the perceptual task was switched to require contrast judgments. In contrast to previous results comparing perception and long-term memory, which have largely provided evidence for domain-specific metacognitive processes, the current findings suggest that metacognition of visual perception and VSTM is supported by a domain-general metacognitive architecture, but only when both domains share the same task-relevant stimulus feature.


2018 ◽  
Vol 74 (8) ◽  
pp. 1298-1307 ◽  
Author(s):  
Patrick S R Davidson ◽  
Petar Vidjen ◽  
Sara Trincao-Batra ◽  
Charles A Collin

Abstract Objectives Pattern separation in memory encoding entails creating and storing distinct, detailed representations to facilitate storage and retrieval. The Mnemonic Similarity Task (MST; Stark, S. M., Yassa, M. A., Lacy, J. W., & Stark, C. E. [2013]. A task to assess behavioral pattern separation [BPS] in humans: Data from healthy aging and mild cognitive impairment. Neuropsychologia, 51, 2442–2449) has been used to argue that normal aging leads to pattern separation decline. We sought to replicate previous reports of age-related difficulty on this behavioral pattern separation estimate and to examine its neuropsychological correlates, specifically long-term memory function, executive function, and visual perception. Methods We administered an object version of the MST to 31 young adults and 38 older adults. It involved a single-probe recognition memory test in which some of the originally studied objects had been replaced with perceptually similar lures, and participants had to identify each as old, a lure, or new. Results Despite their corrected item recognition scores being superior to those of the young adults, the older adults had significantly greater difficulty than the young in discriminating the similar-looking lures from the original items. Interestingly, this lure discrimination difficulty was significantly correlated with visual perception rather than with long-term memory or executive function. Discussion These results suggest that although adult age differences on the MST are reliable, care should be taken to separate perceptual from memory discrimination difficulties as the reason.


1997 ◽  
Vol 9 (6) ◽  
pp. 734-742 ◽  
Author(s):  
Alumit Ishai ◽  
Dov Sagi

Visual imagery and perception share several functional properties and apparently share common underlying brain structures. A main approach to the scientific study of visual imagery is exploring the effects of mental imagery on perceptual processes. Previous studies have shown that visual imagery interferes with perception (Perky effect). Recently we have shown a direct facilitatory effect of visual imagery on visual perception. In an attempt to differentiate the conditions under which visual imagery interferes or facilitates visual perception, we designed new experimental paradigms, using detection tasks of a Gabor target. We found that imagery-induced interference and facilitation are memorydependent: Visual recall of common objects from long-term memory can interfere with perception, while on short-term memory tasks facilitation can be obtained. These results support the distinction between low-level and structural representations in visual memory.


2013 ◽  
Vol 13 (9) ◽  
pp. 926-926
Author(s):  
M. F. Bonner ◽  
A. R. Price ◽  
J. E. Peelle ◽  
M. Grossman

2017 ◽  
Author(s):  
Jason Samaha ◽  
Bradley R Postle

Adaptive behavior depends on the ability to accurately introspect about one’s own performance. Whether this metacognitive ability is supported by the same mechanisms across different tasks is an open question that has thus far been investigated with a focus on correlating metacognitive accuracy between perceptual and long-term memory paradigms. Here, we investigated the relationship between metacognition of visual perception and metacognition of visual short-term memory (VSTM), a cognitive function thought to be more intimately related to visual processing. Across two experiments that required subjects to estimate the perceived or remembered orientation of a grating stimulus and rate their confidence, we observed strong positive correlations between individual differences in metacognitive accuracy between the two tasks. This relationship was not accounted for by individual differences in task performance or average confidence, and was present across two different metrics of metacognition and in both experiments. In contrast to previous results comparing perception and long-term memory, which have largely provided evidence for domain-specific metacognitive processes, the current findings suggest that metacognition of visual perception and VSTM is supported by a domain-general metacognitive architecture.


Sensors ◽  
2021 ◽  
Vol 21 (8) ◽  
pp. 2864
Author(s):  
Yuanping Zhang ◽  
Xiumei Huang ◽  
Ming Yang

To meet the challenge of video target tracking, based on a self-organization mapping network (SOM) and correlation filter, a long-term visual tracking algorithm is proposed. Objects in different videos or images often have completely different appearance, therefore, the self-organization mapping neural network with the characteristics of signal processing mechanism of human brain neurons is used to perform adaptive and unsupervised features learning. A reliable method of robust target tracking is proposed, based on multiple adaptive correlation filters with a memory function of target appearance at the same time. Filters in our method have different updating strategies and can carry out long-term tracking cooperatively. The first is the displacement filter, a kernelized correlation filter that combines contextual characteristics to precisely locate and track targets. Secondly, the scale filters are used to predict the changing scale of a target. Finally, the memory filter is used to maintain the appearance of the target in long-term memory and judge whether the target has failed to track. If the tracking fails, the incremental learning detector is used to recover the target tracking in the way of sliding window. Several experiments show that our method can effectively solve the tracking problems such as severe occlusion, target loss and scale change, and is superior to the state-of-the-art methods in the aspects of efficiency, accuracy and robustness.


2016 ◽  
Vol 39 ◽  
Author(s):  
Mary C. Potter

AbstractRapid serial visual presentation (RSVP) of words or pictured scenes provides evidence for a large-capacity conceptual short-term memory (CSTM) that momentarily provides rich associated material from long-term memory, permitting rapid chunking (Potter 1993; 2009; 2012). In perception of scenes as well as language comprehension, we make use of knowledge that briefly exceeds the supposed limits of working memory.


2020 ◽  
Vol 29 (4) ◽  
pp. 710-727
Author(s):  
Beula M. Magimairaj ◽  
Naveen K. Nagaraj ◽  
Alexander V. Sergeev ◽  
Natalie J. Benafield

Objectives School-age children with and without parent-reported listening difficulties (LiD) were compared on auditory processing, language, memory, and attention abilities. The objective was to extend what is known so far in the literature about children with LiD by using multiple measures and selective novel measures across the above areas. Design Twenty-six children who were reported by their parents as having LiD and 26 age-matched typically developing children completed clinical tests of auditory processing and multiple measures of language, attention, and memory. All children had normal-range pure-tone hearing thresholds bilaterally. Group differences were examined. Results In addition to significantly poorer speech-perception-in-noise scores, children with LiD had reduced speed and accuracy of word retrieval from long-term memory, poorer short-term memory, sentence recall, and inferencing ability. Statistically significant group differences were of moderate effect size; however, standard test scores of children with LiD were not clinically poor. No statistically significant group differences were observed in attention, working memory capacity, vocabulary, and nonverbal IQ. Conclusions Mild signal-to-noise ratio loss, as reflected by the group mean of children with LiD, supported the children's functional listening problems. In addition, children's relative weakness in select areas of language performance, short-term memory, and long-term memory lexical retrieval speed and accuracy added to previous research on evidence-based areas that need to be evaluated in children with LiD who almost always have heterogenous profiles. Importantly, the functional difficulties faced by children with LiD in relation to their test results indicated, to some extent, that commonly used assessments may not be adequately capturing the children's listening challenges. Supplemental Material https://doi.org/10.23641/asha.12808607


Sign in / Sign up

Export Citation Format

Share Document