Stacked Multimodal Attention Network for Context-Aware Video Captioning

Author(s):  
Yi Zheng ◽  
Yuejie Zhang ◽  
Rui Feng ◽  
Tao Zhang ◽  
Weiguo Fan
Author(s):  
Jincan Deng ◽  
Liang Li ◽  
Beichen Zhang ◽  
Shuhui Wang ◽  
Zhengjun Zha ◽  
...  

2019 ◽  
Vol 31 (12) ◽  
pp. 9295-9305 ◽  
Author(s):  
Jiaxu Leng ◽  
Ying Liu ◽  
Shang Chen

2021 ◽  
Author(s):  
Shuqin Chen ◽  
Xian Zhong ◽  
Shifeng Wu ◽  
Zhixin Sun ◽  
Wenxuan Liu ◽  
...  

Author(s):  
Liang Sun ◽  
Bing Li ◽  
Chunfeng Yuan ◽  
Zhengjun Zha ◽  
Weiming Hu

2021 ◽  
Vol 10 (5) ◽  
pp. 336
Author(s):  
Jian Yu ◽  
Meng Zhou ◽  
Xin Wang ◽  
Guoliang Pu ◽  
Chengqi Cheng ◽  
...  

Forecasting the motion of surrounding vehicles is necessary for an autonomous driving system applied in complex traffic. Trajectory prediction helps vehicles make more sensible decisions, which provides vehicles with foresight. However, traditional models consider the trajectory prediction as a simple sequence prediction task. The ignorance of inter-vehicle interaction and environment influence degrades these models in real-world datasets. To address this issue, we propose a novel Dynamic and Static Context-aware Attention Network named DSCAN in this paper. The DSCAN utilizes an attention mechanism to dynamically decide which surrounding vehicles are more important at the moment. We also equip the DSCAN with a constraint network to consider the static environment information. We conducted a series of experiments on a real-world dataset, and the experimental results demonstrated the effectiveness of our model. Moreover, the present study suggests that the attention mechanism and static constraints enhance the prediction results.


Sign in / Sign up

Export Citation Format

Share Document