Attachment anxiety moderates the effect of oxytocin on negative emotion recognition: Evidence from eye-movement data

2020 ◽  
Vol 198 ◽  
pp. 173015
Author(s):  
Tianyu Wang ◽  
Qingting Tang ◽  
Xin Wu ◽  
Xu Chen
2022 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Fatima Isiaka ◽  
Zainab Adamu

PurposeOne of the contributions of artificial intelligent (AI) in modern technology is emotion recognition which is mostly based on facial expression and modification of its inference engine. The facial recognition scheme is mostly built to understand user expression in an online business webpage on a marketing site but has limited abilities to recognise elusive expressions. The basic emotions are expressed when interrelating and socialising with other personnel online. At most times, studying how to understand user expression is often a most tedious task, especially the subtle expressions. An emotion recognition system can be used to optimise and reduce complexity in understanding users' subconscious thoughts and reasoning through their pupil changes.Design/methodology/approachThis paper demonstrates the use of personal computer (PC) webcam to read in eye movement data that includes pupil changes as part of distinct user attributes. A custom eye movement algorithm (CEMA) is used to capture users' activity and record the data which is served as an input model to an inference engine (artificial neural network (ANN)) that helps to predict user emotional response conveyed as emoticons on the webpage.FindingsThe result from the error in performance shows that ANN is most adaptable to user behaviour prediction and can be used for the system's modification paradigm.Research limitations/implicationsOne of the drawbacks of the analytical tool is its inability in some cases to set some of the emoticons within the boundaries of the visual field, this is a limitation to be tackled within subsequent runs with standard techniques.Originality/valueThe originality of the proposed model is its ability to predict basic user emotional response based on changes in pupil size between average recorded baseline boundaries and convey the emoticons chronologically with the gaze points.


2019 ◽  
Vol 24 (4) ◽  
pp. 297-311
Author(s):  
José David Moreno ◽  
José A. León ◽  
Lorena A. M. Arnal ◽  
Juan Botella

Abstract. We report the results of a meta-analysis of 22 experiments comparing the eye movement data obtained from young ( Mage = 21 years) and old ( Mage = 73 years) readers. The data included six eye movement measures (mean gaze duration, mean fixation duration, total sentence reading time, mean number of fixations, mean number of regressions, and mean length of progressive saccade eye movements). Estimates were obtained of the typified mean difference, d, between the age groups in all six measures. The results showed positive combined effect size estimates in favor of the young adult group (between 0.54 and 3.66 in all measures), although the difference for the mean number of fixations was not significant. Young adults make in a systematic way, shorter gazes, fewer regressions, and shorter saccadic movements during reading than older adults, and they also read faster. The meta-analysis results confirm statistically the most common patterns observed in previous research; therefore, eye movements seem to be a useful tool to measure behavioral changes due to the aging process. Moreover, these results do not allow us to discard either of the two main hypotheses assessed for explaining the observed aging effects, namely neural degenerative problems and the adoption of compensatory strategies.


2014 ◽  
Author(s):  
Bernhard Angele ◽  
Elizabeth R. Schotter ◽  
Timothy Slattery ◽  
Tara L. Chaloukian ◽  
Klinton Bicknell ◽  
...  

Author(s):  
Ayush Kumar ◽  
Prantik Howlader ◽  
Rafael Garcia ◽  
Daniel Weiskopf ◽  
Klaus Mueller

2021 ◽  
Vol 1757 (1) ◽  
pp. 012021
Author(s):  
Yuqiong Wang ◽  
Zehui Zhao ◽  
Zhiwei Huang

Sensors ◽  
2021 ◽  
Vol 21 (15) ◽  
pp. 5178
Author(s):  
Sangbong Yoo ◽  
Seongmin Jeong ◽  
Seokyeon Kim ◽  
Yun Jang

Gaze movement and visual stimuli have been utilized to analyze human visual attention intuitively. Gaze behavior studies mainly show statistical analyses of eye movements and human visual attention. During these analyses, eye movement data and the saliency map are presented to the analysts as separate views or merged views. However, the analysts become frustrated when they need to memorize all of the separate views or when the eye movements obscure the saliency map in the merged views. Therefore, it is not easy to analyze how visual stimuli affect gaze movements since existing techniques focus excessively on the eye movement data. In this paper, we propose a novel visualization technique for analyzing gaze behavior using saliency features as visual clues to express the visual attention of an observer. The visual clues that represent visual attention are analyzed to reveal which saliency features are prominent for the visual stimulus analysis. We visualize the gaze data with the saliency features to interpret the visual attention. We analyze the gaze behavior with the proposed visualization to evaluate that our approach to embedding saliency features within the visualization supports us to understand the visual attention of an observer.


1972 ◽  
Vol 35 (1) ◽  
pp. 103-110
Author(s):  
Phillip Kleespies ◽  
Morton Wiener

This study explored (1) for evidence of visual input at so-called “subliminal” exposure durations, and (2) whether the response, if any, was a function of the thematic content of the stimulus. Thematic content (threatening versus non-threatening) and stimulus structure (angular versus curved) were varied independently under “subliminal,” “part-cue,” and “identification” exposure conditions. With Ss' reports and the frequency and latency of first eye movements (“orienting reflex”) as input indicators, there was no evidence of input differences which are a function of thematic content at any exposure duration, and the “report” data were consistent with the eye-movement data.


2015 ◽  
Vol 68 ◽  
pp. 158-167 ◽  
Author(s):  
Sandra Baez ◽  
Eduar Herrera ◽  
Oscar Gershanik ◽  
Adolfo M. Garcia ◽  
Yamile Bocanegra ◽  
...  

Array ◽  
2021 ◽  
pp. 100087
Author(s):  
Peter Raatikainen ◽  
Jarkko Hautala ◽  
Otto Loberg ◽  
Tommi Kärkkäinen ◽  
Paavo Leppänen ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document