Classification of Optical Flow by Constraints

Author(s):  
Yusuke Kameda ◽  
Atsushi Imiya
Keyword(s):  
2020 ◽  
Vol 17 (4) ◽  
pp. 497-506
Author(s):  
Sunil Patel ◽  
Ramji Makwana

Automatic classification of dynamic hand gesture is challenging due to the large diversity in a different class of gesture, Low resolution, and it is performed by finger. Due to a number of challenges many researchers focus on this area. Recently deep neural network can be used for implicit feature extraction and Soft Max layer is used for classification. In this paper, we propose a method based on a two-dimensional convolutional neural network that performs detection and classification of hand gesture simultaneously from multimodal Red, Green, Blue, Depth (RGBD) and Optical flow Data and passes this feature to Long-Short Term Memory (LSTM) recurrent network for frame-to-frame probability generation with Connectionist Temporal Classification (CTC) network for loss calculation. We have calculated an optical flow from Red, Green, Blue (RGB) data for getting proper motion information present in the video. CTC model is used to efficiently evaluate all possible alignment of hand gesture via dynamic programming and check consistency via frame-to-frame for the visual similarity of hand gesture in the unsegmented input stream. CTC network finds the most probable sequence of a frame for a class of gesture. The frame with the highest probability value is selected from the CTC network by max decoding. This entire CTC network is trained end-to-end with calculating CTC loss for recognition of the gesture. We have used challenging Vision for Intelligent Vehicles and Applications (VIVA) dataset for dynamic hand gesture recognition captured with RGB and Depth data. On this VIVA dataset, our proposed hand gesture recognition technique outperforms competing state-of-the-art algorithms and gets an accuracy of 86%


Action recognition (AR) plays a fundamental role in computer vision and video analysis. We are witnessing an astronomical increase of video data on the web and it is difficult to recognize the action in video due to different view point of camera. For AR in video sequence, it depends upon appearance in frame and optical flow in frames of video. In video spatial and temporal components of video frames features play integral role for better classification of action in videos. In the proposed system, RGB frames and optical flow frames are used for AR with the help of Convolutional Neural Network (CNN) pre-trained model Alex-Net extract features from fc7 layer. Support vector machine (SVM) classifier is used for the classification of AR in videos. For classification purpose, HMDB51 dataset have been used which includes 51 Classes of human action. The dataset is divided into 51 action categories. Using SVM classifier, extracted features are used for classification and achieved best result 95.6% accuracy as compared to other techniques of the state-of- art.v


2020 ◽  
Vol 148 (4) ◽  
pp. 2655-2655
Author(s):  
Matthew Faytak ◽  
Connor Mayer ◽  
Jennifer Kuo ◽  
G. Teixeira ◽  
Z. L. Zhou

2021 ◽  
Vol 7 (5) ◽  
pp. 90
Author(s):  
Slim Hamdi ◽  
Samir Bouindour ◽  
Hichem Snoussi ◽  
Tian Wang ◽  
Mohamed Abid

In recent years, the use of drones for surveillance tasks has been on the rise worldwide. However, in the context of anomaly detection, only normal events are available for the learning process. Therefore, the implementation of a generative learning method in an unsupervised mode to solve this problem becomes fundamental. In this context, we propose a new end-to-end architecture capable of generating optical flow images from original UAV images and extracting compact spatio-temporal characteristics for anomaly detection purposes. It is designed with a custom loss function as a sum of three terms, the reconstruction loss (Rl), the generation loss (Gl) and the compactness loss (Cl) to ensure an efficient classification of the “deep-one” class. In addition, we propose to minimize the effect of UAV motion in video processing by applying background subtraction on optical flow images. We tested our method on very complex datasets called the mini-drone video dataset, and obtained results surpassing existing techniques’ performances with an AUC of 85.3.


Author(s):  
V. V. Kniaz ◽  
Z. N. Smirnova

Human emotion identification from image sequences is highly demanded nowadays. The range of possible applications can vary from an automatic smile shutter function of consumer grade digital cameras to Biofied Building technologies, which enables communication between building space and residents. The highly perceptual nature of human emotions leads to the complexity of their classification and identification. The main question arises from the subjective quality of emotional classification of events that elicit human emotions. A variety of methods for formal classification of emotions were developed in musical psychology. This work is focused on identification of human emotions evoked by musical pieces using human face tracking and optical flow analysis. Facial feature tracking algorithm used for facial feature speed and position estimation is presented. <br><br> Facial features were extracted from each image sequence using human face tracking with local binary patterns (LBP) features. Accurate relative speeds of facial features were estimated using optical flow analysis. Obtained relative positions and speeds were used as the output facial emotion vector. The algorithm was tested using original software and recorded image sequences. The proposed technique proves to give a robust identification of human emotions elicited by musical pieces. The estimated models could be used for human emotion identification from image sequences in such fields as emotion based musical background or mood dependent radio.


2021 ◽  
Vol 33 (2) ◽  
pp. 292-300
Author(s):  
Chihiro Kikumoto ◽  
Yoh Harimoto ◽  
Kazuki Isogaya ◽  
Takeshi Yoshida ◽  
Takateru Urakubo ◽  
...  

The increased use of UAVs (Unmanned Aerial Vehicles) has heightened demands for an automated landing system intended for a variety of tasks and emergency landings. A key challenge of this system is finding a safe landing site in an unknown environment using on-board sensors. This paper proposes a method to generate a heat map for safety evaluation using images from a single on-board camera. The proposed method consists of the classification of ground surface by CNNs (Convolutional Neural Networks) and the estimation of surface flatness from optical flow. We present the results of applying this method to a video obtained from an on-board camera and discuss ways of improving the method.


2019 ◽  
Vol 102 (9) ◽  
pp. 19-26 ◽  
Author(s):  
Yutaka Suzuki ◽  
Kyosuke Hatsushika ◽  
Keisuke Masuyama ◽  
Osamu Sakata ◽  
Morimasa Tanimoto ◽  
...  

1966 ◽  
Vol 24 ◽  
pp. 21-23
Author(s):  
Y. Fujita

We have investigated the spectrograms (dispersion: 8Å/mm) in the photographic infrared region fromλ7500 toλ9000 of some carbon stars obtained by the coudé spectrograph of the 74-inch reflector attached to the Okayama Astrophysical Observatory. The names of the stars investigated are listed in Table 1.


Author(s):  
Gerald Fine ◽  
Azorides R. Morales

For years the separation of carcinoma and sarcoma and the subclassification of sarcomas has been based on the appearance of the tumor cells and their microscopic growth pattern and information derived from certain histochemical and special stains. Although this method of study has produced good agreement among pathologists in the separation of carcinoma from sarcoma, it has given less uniform results in the subclassification of sarcomas. There remain examples of neoplasms of different histogenesis, the classification of which is questionable because of similar cytologic and growth patterns at the light microscopic level; i.e. amelanotic melanoma versus carcinoma and occasionally sarcoma, sarcomas with an epithelial pattern of growth simulating carcinoma, histologically similar mesenchymal tumors of different histogenesis (histiocytoma versus rhabdomyosarcoma, lytic osteogenic sarcoma versus rhabdomyosarcoma), and myxomatous mesenchymal tumors of diverse histogenesis (myxoid rhabdo and liposarcomas, cardiac myxoma, myxoid neurofibroma, etc.)


Sign in / Sign up

Export Citation Format

Share Document