temporal action
Recently Published Documents


TOTAL DOCUMENTS

254
(FIVE YEARS 188)

H-INDEX

21
(FIVE YEARS 7)

2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Quanping Shen ◽  
Songzhong Ye

Technical movement analysis requires specialized domain knowledge and processing a large amount of data, and the advantages of AI in processing data can improve the efficiency of data analysis. In this paper, we propose a feature pyramid network-based temporal action detection (FPN-TAD) algorithm, which is used to solve the problem that the action proposal module has a low recall rate for small-scale temporal target action regions in the current video temporal action detection algorithm research. This paper is divided into three parts. The first part is an overview of the algorithm; the second part elaborates the network structure and the working principle of the FPN-TAD algorithm; and the third part gives the experimental results and analysis of the algorithm.


Sensors ◽  
2021 ◽  
Vol 21 (24) ◽  
pp. 8309
Author(s):  
Inwoong Lee ◽  
Doyoung Kim ◽  
Dongyoon Wee ◽  
Sanghoon Lee

In recent years, human action recognition has been studied by many computer vision researchers. Recent studies have attempted to use two-stream networks using appearance and motion features, but most of these approaches focused on clip-level video action recognition. In contrast to traditional methods which generally used entire images, we propose a new human instance-level video action recognition framework. In this framework, we represent the instance-level features using human boxes and keypoints, and our action region features are used as the inputs of the temporal action head network, which makes our framework more discriminative. We also propose novel temporal action head networks consisting of various modules, which reflect various temporal dynamics well. In the experiment, the proposed models achieve comparable performance with the state-of-the-art approaches on two challenging datasets. Furthermore, we evaluate the proposed features and networks to verify the effectiveness of them. Finally, we analyze the confusion matrix and visualize the recognized actions at human instance level when there are several people.


2021 ◽  
Vol 16 (4) ◽  
Author(s):  
Tian Wang ◽  
Shiye Lei ◽  
Youyou Jiang ◽  
Choi Chang ◽  
Hichem Snoussi ◽  
...  

2021 ◽  
Author(s):  
Tingting Han ◽  
Sicheng Zhao ◽  
Xiaoshuai Sun ◽  
Jun Yu
Keyword(s):  

2021 ◽  
Author(s):  
Morgan Liang ◽  
Xun Li ◽  
Sandersan Onie ◽  
Mark Larsen ◽  
Arcot Sowmya

2021 ◽  
Vol 12 (5) ◽  
pp. 1-14
Author(s):  
Yisheng Zhu ◽  
Hu Han ◽  
Guangcan Liu ◽  
Qingshan Liu

Temporal action proposal generation is an essential and challenging task in video understanding, which aims to locate the temporal intervals that likely contain the actions of interest. Although great progress has been made, the problem is still far from being well solved. In particular, prevalent methods can handle well only the local dependencies (i.e., short-term dependencies) among adjacent frames but are generally powerless in dealing with the global dependencies (i.e., long-term dependencies) between distant frames. To tackle this issue, we propose CLGNet, a novel Collaborative Local-Global Learning Network for temporal action proposal. The majority of CLGNet is an integration of Temporal Convolution Network and Bidirectional Long Short-Term Memory, in which Temporal Convolution Network is responsible for local dependencies while Bidirectional Long Short-Term Memory takes charge of handling the global dependencies. Furthermore, an attention mechanism called the background suppression module is designed to guide our model to focus more on the actions. Extensive experiments on two benchmark datasets, THUMOS’14 and ActivityNet-1.3, show that the proposed method can outperform state-of-the-art methods, demonstrating the strong capability of modeling the actions with varying temporal durations.


Sign in / Sign up

Export Citation Format

Share Document