Visual Question Answering using Explicit Visual Attention

Author(s):  
Vasileios Lioutas ◽  
Nikolaos Passalis ◽  
Anastasios Tefas
Author(s):  
Jingkuan Song ◽  
Pengpeng Zeng ◽  
Lianli Gao ◽  
Heng Tao Shen

Recently, attention-based Visual Question Answering (VQA) has achieved great success by utilizing question to selectively target different visual areas that are related to the answer. Existing visual attention models are generally planar, i.e., different channels of the last conv-layer feature map of an image share the same weight. This conflicts with the attention mechanism because CNN features are naturally spatial and channel-wise. Also, visual attention models are usually conducted on pixel-level, which may cause region discontinuous problem. In this paper we propose a Cubic Visual Attention (CVA) model by successfully applying a novel channel and spatial attention on object regions to improve VQA task. Specifically, instead of attending to pixels, we first take advantage of the object proposal networks to generate a set of object candidates and extract their associated conv features. Then, we utilize the question to guide channel attention and spatial attention calculation based on the con-layer feature map. Finally, the attended visual features and the question are combined to infer the answer. We assess the performance of our proposed CVA on three public image QA datasets, including COCO-QA, VQA and Visual7W. Experimental results show that our proposed method significantly outperforms the state-of-the-arts.


2021 ◽  
Author(s):  
Dezhi Han ◽  
Shuli Zhou ◽  
Kuan Ching Li ◽  
Rodrigo Fernandes de Mello

2021 ◽  
Vol 1828 (1) ◽  
pp. 012145
Author(s):  
Ye Qin ◽  
Zhiping Zhou ◽  
Chen Biao ◽  
Li Wenjie

Sign in / Sign up

Export Citation Format

Share Document