Pedestrian violence detection based on optical flow energy characteristics

Author(s):  
Zhonghua Guo ◽  
Fengjie Wu ◽  
Haishan Chen ◽  
Junying Yuan ◽  
Canzeng Cai
2013 ◽  
Author(s):  
Zhijie Yang ◽  
Tao Zhang ◽  
Jie Yang ◽  
Qiang Wu ◽  
Li Bai ◽  
...  

2018 ◽  
Vol 14 (04) ◽  
pp. 193 ◽  
Author(s):  
Honghua Xu ◽  
Li Li ◽  
Ming Fang ◽  
Fengrong Zhang

In this paper, the main technologies of foreground detection, feature description and extraction, movement behavior classification and recognition were introduced. Based on optical flow for movement objects detection, optical flow energy image was put forward for movement feature expression and region convolutional neural networks was adopt to choose features and decrease dimension. Then support vector machine classifier was trained and used to classify and recognize actions. After training and testing on public human actions database, the experiment result showed that the method could effectively distinguish human actions and significantly improved the recognition accuracy of human actions. And for the different situations of camera lens drawing near, pulling away or slight movement of camera, the solution had recognition effect as well. At last, this scheme was applied to intelligent video surveillance system, which was used to identify abnormal behavior and alarm. The abnormal behaviors of faint, smashing car, robbery and fighting were defined in the system. In running of the system, it obtained satisfactory recognition results.


Electronics ◽  
2021 ◽  
Vol 10 (13) ◽  
pp. 1601
Author(s):  
Fernando J. Rendón-Segador ◽  
Juan A. Álvarez-García ◽  
Fernando Enríquez ◽  
Oscar Deniz

Introducing efficient automatic violence detection in video surveillance or audiovisual content monitoring systems would greatly facilitate the work of closed-circuit television (CCTV) operators, rating agencies or those in charge of monitoring social network content. In this paper we present a new deep learning architecture, using an adapted version of DenseNet for three dimensions, a multi-head self-attention layer and a bidirectional convolutional long short-term memory (LSTM) module, that allows encoding relevant spatio-temporal features, to determine whether a video is violent or not. Furthermore, an ablation study of the input frames, comparing dense optical flow and adjacent frames subtraction and the influence of the attention layer is carried out, showing that the combination of optical flow and the attention mechanism improves results up to 4.4%. The conducted experiments using four of the most widely used datasets for this problem, matching or exceeding in some cases the results of the state of the art, reducing the number of network parameters needed (4.5 millions), and increasing its efficiency in test accuracy (from 95.6% on the most complex dataset to 100% on the simplest one) and inference time (less than 0.3 s for the longest clips). Finally, to check if the generated model is able to generalize violence, a cross-dataset analysis is performed, which shows the complexity of this approach: using three datasets to train and testing on the remaining one the accuracy drops in the worst case to 70.08% and in the best case to 81.51%, which points to future work oriented towards anomaly detection in new datasets.


Sign in / Sign up

Export Citation Format

Share Document