scholarly journals Multi-Level Feature Abstraction from Convolutional Neural Networks for Multimodal Biometric Identification

Author(s):  
Sobhan Soleymani ◽  
Ali Dabouei ◽  
Hadi Kazemi ◽  
Jeremy Dawson ◽  
Nasser M. Nasrabadi
2020 ◽  
Vol 391 ◽  
pp. 83-95 ◽  
Author(s):  
Yazhao Li ◽  
Yanwei Pang ◽  
Kongqiao Wang ◽  
Xuelong Li

IEEE Access ◽  
2019 ◽  
Vol 7 ◽  
pp. 74973-74985 ◽  
Author(s):  
Pengju Liu ◽  
Hongzhi Zhang ◽  
Wei Lian ◽  
Wangmeng Zuo

2020 ◽  
Vol 34 (07) ◽  
pp. 12410-12417 ◽  
Author(s):  
Xinyi Wu ◽  
Zhenyao Wu ◽  
Jinglin Zhang ◽  
Lili Ju ◽  
Song Wang

The performance of predicting human fixations in videos has been much enhanced with the help of development of the convolutional neural networks (CNN). In this paper, we propose a novel end-to-end neural network “SalSAC” for video saliency prediction, which uses the CNN-LSTM-Attention as the basic architecture and utilizes the information from both static and dynamic aspects. To better represent the static information of each frame, we first extract multi-level features of same size from different layers of the encoder CNN and calculate the corresponding multi-level attentions, then we randomly shuffle these attention maps among levels and multiply them to the extracted multi-level features respectively. Through this way, we leverage the attention consistency across different layers to improve the robustness of the network. On the dynamic aspect, we propose a correlation-based ConvLSTM to appropriately balance the influence of the current and preceding frames to the prediction. Experimental results on the DHF1K, Hollywood2 and UCF-sports datasets show that SalSAC outperforms many existing state-of-the-art methods.


Sign in / Sign up

Export Citation Format

Share Document