Multi-window visual saliency extraction for fusion of visible and infrared images

2016 ◽  
Vol 76 ◽  
pp. 295-302 ◽  
Author(s):  
Jufeng Zhao ◽  
Xiumin Gao ◽  
Yueting Chen ◽  
Huajun Feng ◽  
Daodang Wang
Author(s):  
Snehal S. Rajole ◽  
J. V. Shinde

In this paper we proposed unique technique which is adaptive to noisy images for eye gaze detection as processing noisy sclera images captured at-a-distance and on-the-move has not been extensively investigated. Sclera blood vessels have been investigated recently as an efficient biometric trait. Capturing part of the eye with a normal camera using visible-wavelength images rather than near infrared images has provoked research interest. This technique involves sclera template rotation alignment and a distance scaling method to minimize the error rates when noisy eye images are captured at-a-distance and on-the move. The proposed system is tested and results are generated by extensive simulation in java.


2017 ◽  
Vol 9 (3) ◽  
pp. 235-240
Author(s):  
Shimoga N. B. Bhushan ◽  
. Harisha ◽  
Arti Pawar ◽  
. Vidyalakshmi

1999 ◽  
Vol 117 (1) ◽  
pp. 439-445 ◽  
Author(s):  
P. Persi ◽  
A. R. Marenzi ◽  
A. A. Kaas ◽  
G. Olofsson ◽  
L. Nordh ◽  
...  

2018 ◽  
Vol 1098 ◽  
pp. 012033
Author(s):  
Ying Lin ◽  
Jiafeng Qin ◽  
Weiwei Zhang ◽  
Hao Zhang ◽  
Demeng Bai ◽  
...  

2021 ◽  
Vol 11 (16) ◽  
pp. 7217
Author(s):  
Cristina Luna-Jiménez ◽  
Jorge Cristóbal-Martín ◽  
Ricardo Kleinlein ◽  
Manuel Gil-Martín ◽  
José M. Moya ◽  
...  

Spatial Transformer Networks are considered a powerful algorithm to learn the main areas of an image, but still, they could be more efficient by receiving images with embedded expert knowledge. This paper aims to improve the performance of conventional Spatial Transformers when applied to Facial Expression Recognition. Based on the Spatial Transformers’ capacity of spatial manipulation within networks, we propose different extensions to these models where effective attentional regions are captured employing facial landmarks or facial visual saliency maps. This specific attentional information is then hardcoded to guide the Spatial Transformers to learn the spatial transformations that best fit the proposed regions for better recognition results. For this study, we use two datasets: AffectNet and FER-2013. For AffectNet, we achieve a 0.35% point absolute improvement relative to the traditional Spatial Transformer, whereas for FER-2013, our solution gets an increase of 1.49% when models are fine-tuned with the Affectnet pre-trained weights.


2021 ◽  
Author(s):  
Sai Phani Kumar Malladi ◽  
Jayanta Mukhopadhyay ◽  
Chaker Larabi ◽  
Santanu Chaudhury

Author(s):  
Shenyi Qian ◽  
Yongsheng Shi ◽  
Huaiguang Wu ◽  
Jinhua Liu ◽  
Weiwei Zhang

Sign in / Sign up

Export Citation Format

Share Document