Leveraging auxiliary image descriptions for dense video captioning

2021 ◽  
Vol 146 ◽  
pp. 70-76
Author(s):  
Emre Boran ◽  
Aykut Erdem ◽  
Nazli Ikizler-Cinbis ◽  
Erkut Erdem ◽  
Pranava Madhyastha ◽  
...  
2019 ◽  
Vol 23 (1) ◽  
pp. 147-159
Author(s):  
Shagan Sah ◽  
Thang Nguyen ◽  
Ray Ptucha

Author(s):  
Alok Singh ◽  
Thoudam Doren Singh ◽  
Sivaji Bandyopadhyay
Keyword(s):  

Author(s):  
Jincan Deng ◽  
Liang Li ◽  
Beichen Zhang ◽  
Shuhui Wang ◽  
Zhengjun Zha ◽  
...  

2021 ◽  
Vol 7 (2) ◽  
pp. 12
Author(s):  
Yousef I. Mohamad ◽  
Samah S. Baraheem ◽  
Tam V. Nguyen

Automatic event recognition in sports photos is both an interesting and valuable research topic in the field of computer vision and deep learning. With the rapid increase and the explosive spread of data, which is being captured momentarily, the need for fast and precise access to the right information has become a challenging task with considerable importance for multiple practical applications, i.e., sports image and video search, sport data analysis, healthcare monitoring applications, monitoring and surveillance systems for indoor and outdoor activities, and video captioning. In this paper, we evaluate different deep learning models in recognizing and interpreting the sport events in the Olympic Games. To this end, we collect a dataset dubbed Olympic Games Event Image Dataset (OGED) including 10 different sport events scheduled for the Olympic Games Tokyo 2020. Then, the transfer learning is applied on three popular deep convolutional neural network architectures, namely, AlexNet, VGG-16 and ResNet-50 along with various data augmentation methods. Extensive experiments show that ResNet-50 with the proposed photobombing guided data augmentation achieves 90% in terms of accuracy.


2015 ◽  
Vol 7 (4) ◽  
pp. 1-21 ◽  
Author(s):  
Valerie S. Morash ◽  
Yue-Ting Siu ◽  
Joshua A. Miele ◽  
Lucia Hasty ◽  
Steven Landau
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document