video descriptor
Recently Published Documents


TOTAL DOCUMENTS

10
(FIVE YEARS 4)

H-INDEX

3
(FIVE YEARS 2)

2019 ◽  
Vol 79 (9-10) ◽  
pp. 6025-6043 ◽  
Author(s):  
Mehrin Saremi ◽  
Farzin Yaghmaee

2019 ◽  
Author(s):  
Matheus Vieira Lessa Ribeiro ◽  
Jorge Leonid Aching Samatelo

Traffic congestion is a significant problem in urban cities and affects economic, health, and social questions. Although many works have been published in the last years to traffic applications based on video data, different techniques of computer vision can be explored in this area. In this work, we proposed a method for traffic flow classification using StarRGB and Convolutional Neural Networks (CNN). The StarRGB describes a global representation of the traffic video into a colored image based on motion elements in the scene. Then, the generated image passed as input to a pre-trained CNN to extract the features and classify the traffic video activity in three classes: LIGHT, MEDIUM, and HEAVY. In our experiments using a traffic video database, the proposed method reached an accuracy of 96.47%. Also, the results suggest that StarRGB is a good descriptor for traffic video applications.


2019 ◽  
Vol 15 (29) ◽  
pp. 82-94
Author(s):  
Fabio Martínez Carrillo ◽  
Fabián Castillo ◽  
Lola Bautista

RGB-D sensors have allowed attacking many classical problems in computer vision such as segmentation, scene representations and human interaction, among many others. Regarding motion characterization, typical RGB-D strategies are limited to namely analyze global shape changes and capture scene flow fields to describe local motions in depth sequences. Nevertheless, such strategies only recover motion information among a couple of frames, limiting the analysis of coherent large displacements along time. This work presents a novel strategy to compute 3D+t dense and long motion trajectories as fundamental kinematic primitives to represent video sequences. Each motion trajectory models kinematic words primitives that together can describe complex gestures developed along videos. Such kinematic words were processed into a bag-of-kinematic-words framework to obtain an occurrence video descriptor. The novel video descriptor based on 3D+t motion trajectories achieved an average accuracy of 80% in a dataset of 5 gestures and 100 videos.


Symmetry ◽  
2019 ◽  
Vol 11 (1) ◽  
pp. 52 ◽  
Author(s):  
Xianzhang Pan ◽  
Wenping Guo ◽  
Xiaoying Guo ◽  
Wenshu Li ◽  
Junjie Xu ◽  
...  

The proposed method has 30 streams, i.e., 15 spatial streams and 15 temporal streams. Each spatial stream corresponds to each temporal stream. Therefore, this work correlates with the symmetry concept. It is a difficult task to classify video-based facial expression owing to the gap between the visual descriptors and the emotions. In order to bridge the gap, a new video descriptor for facial expression recognition is presented to aggregate spatial and temporal convolutional features across the entire extent of a video. The designed framework integrates a state-of-the-art 30 stream and has a trainable spatial–temporal feature aggregation layer. This framework is end-to-end trainable for video-based facial expression recognition. Thus, this framework can effectively avoid overfitting to the limited emotional video datasets, and the trainable strategy can learn to better represent an entire video. The different schemas for pooling spatial–temporal features are investigated, and the spatial and temporal streams are best aggregated by utilizing the proposed method. The extensive experiments on two public databases, BAUM-1s and eNTERFACE05, show that this framework has promising performance and outperforms the state-of-the-art strategies.


2016 ◽  
Vol 16 (04) ◽  
pp. 1650017
Author(s):  
Felipe Andrade Caetano ◽  
Marcelo Bernardes Vieira ◽  
Rodrigo Luis de Souza da Silva

Dense trajectories have been shown as a very promising method in the human action recognition field. In this paper, we propose a new kind of video descriptor, generated from the relationship between the trajectory’s optical flow with the gradient field in its neighborhood. Orientation tensors are used to accumulate relevant information over the video, representing the tendency of direction in the descriptor space for that kind of movement. Furthermore, a method to cluster trajectories using their shape is proposed. This method allows us to accumulate different motion patterns in different tensors and easier distinguish trajectories that are created by real movements from the trajectories created by the camera’s movement. The proposed method is capable to achieve the best known recognition rates for methods based on the self-descriptor constraint in popular datasets — Hollywood2 (up to 46%) and KTH (up to 94%).


2012 ◽  
Vol 24 (7) ◽  
pp. 1473-1485 ◽  
Author(s):  
Berkan Solmaz ◽  
Shayan Modiri Assari ◽  
Mubarak Shah
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document