Feature Extraction of Video Using Artificial Neural Network

Author(s):  
Yoshihiro Hayakawa ◽  
Takanori Oonuma ◽  
Hideyuki Kobayashi ◽  
Akiko Takahashi ◽  
Shinji Chiba ◽  
...  

In deep neural networks, which have been gaining attention in recent years, the features of input images are expressed in a middle layer. Using the information on this feature layer, high performance can be demonstrated in the image recognition field. In the present study, we achieve image recognition, without using convolutional neural networks or sparse coding, through an image feature extraction function obtained when identity mapping learning is applied to sandglass-style feed-forward neural networks. In sports form analysis, for example, a state trajectory is mapped in a low-dimensional feature space based on a consecutive series of actions. Here, we discuss ideas related to image analysis by applying the above method.

Author(s):  
Yoshihiro Hayakawa ◽  
Takanori Oonuma ◽  
Hideyuki Kobayashi ◽  
Akiko Takahashi ◽  
Shinji Chiba ◽  
...  

In deep neural networks, which have been gaining attention in recent years, the features of input images are expressed in a middle layer. Using the information on this feature layer, high performance can be demonstrated in the image recognition field. In the present study, we achieve image recognition, without using convolutional neural networks or sparse coding, through an image feature extraction function obtained when identity mapping learning is applied to sandglass-style feed-forward neural networks. In sports form analysis, for example, a state trajectory is mapped in a low-dimensional feature space based on a consecutive series of actions. Here, we discuss ideas related to image analysis by applying the above method.


2020 ◽  
Vol 2020 ◽  
pp. 1-21 ◽  
Author(s):  
Hong Yang ◽  
Yasheng Zhang ◽  
Wenzhe Ding

Feature extraction is the key step of Inverse Synthetic Aperture Radar (ISAR) image recognition. However, limited by the cost and conditions of ISAR image acquisition, it is relatively difficult to obtain large-scale sample data, which makes it difficult to obtain target deep features with good discriminability by using the currently popular deep learning method. In this paper, a new method for low-dimensional, strongly robust, and fast space target ISAR image recognition based on local and global structural feature fusion is proposed. This method performs the trace transformation along the longest axis of the ISAR image to generate the global trace feature of the space target ISAR image. By introducing the local structural feature, Local Binary Pattern (LBP), the complementary fusion of the global and local features is achieved, which makes up for the missing structural information of the trace feature and ensures the integrity of the ISAR image feature information. The representation of trace and LBP features in a low-dimensional mapping feature space is found by using the manifold learning method. Under the condition of maintaining the local neighborhood relationship in the original feature space, the effective fusion of trace and LBP features is achieved. So, in the practical application process, the target recognition accuracy is no longer affected by trace function, LBP feature block number selection, and other factors, realizing the high robustness of the algorithm. To verify the effectiveness of the proposed algorithm, an ISAR image database containing 1325 samples of 5 types of space targets is used for experiments. The results show that the classification accuracy of the 5 types of space targets can reach more than 99%, and the recognition accuracy is no longer affected by the trace feature and LBP feature selection, which has strong robustness. The proposed method provides a fast and effective high-precision model for space target feature extraction, which can give some references for solving the problem of space object efficient identification under the condition of small sample data.


2021 ◽  
Author(s):  
Rogini Runghen ◽  
Daniel B Stouffer ◽  
Giulio Valentino Dalla Riva

Collecting network interaction data is difficult. Non-exhaustive sampling and complex hidden processes often result in an incomplete data set. Thus, identifying potentially present but unobserved interactions is crucial both in understanding the structure of large scale data, and in predicting how previously unseen elements will interact. Recent studies in network analysis have shown that accounting for metadata (such as node attributes) can improve both our understanding of how nodes interact with one another, and the accuracy of link prediction. However, the dimension of the object we need to learn to predict interactions in a network grows quickly with the number of nodes. Therefore, it becomes computationally and conceptually challenging for large networks. Here, we present a new predictive procedure combining a graph embedding method with machine learning techniques to predict interactions on the base of nodes' metadata. Graph embedding methods project the nodes of a network onto a---low dimensional---latent feature space. The position of the nodes in the latent feature space can then be used to predict interactions between nodes. Learning a mapping of the nodes' metadata to their position in a latent feature space corresponds to a classic---and low dimensional---machine learning problem. In our current study we used the Random Dot Product Graph model to estimate the embedding of an observed network, and we tested different neural networks architectures to predict the position of nodes in the latent feature space. Flexible machine learning techniques to map the nodes onto their latent positions allow to account for multivariate and possibly complex nodes' metadata. To illustrate the utility of the proposed procedure, we apply it to a large dataset of tourist visits to destinations across New Zealand. We found that our procedure accurately predicts interactions for both existing nodes and nodes newly added to the network, while being computationally feasible even for very large networks. Overall, our study highlights that by exploiting the properties of a well understood statistical model for complex networks and combining it with standard machine learning techniques, we can simplify the link prediction problem when incorporating multivariate node metadata. Our procedure can be immediately applied to different types of networks, and to a wide variety of data from different systems. As such, both from a network science and data science perspective, our work offers a flexible and generalisable procedure for link prediction.


2021 ◽  
Vol 2083 (4) ◽  
pp. 042007
Author(s):  
Xiaowen Liu ◽  
Juncheng Lei

Abstract Image recognition technology mainly includes image feature extraction and classification recognition. Feature extraction is the key link, which determines whether the recognition performance is good or bad. Deep learning builds a model by building a hierarchical model structure like the human brain, extracting features layer by layer from the data. Applying deep learning to image recognition can further improve the accuracy of image recognition. Based on the idea of clustering, this article establishes a multi-mix Gaussian model for engineering image information in RGB color space through offline learning and expectation-maximization algorithms, to obtain a multi-mix cluster representation of engineering image information. Then use the sparse Gaussian machine learning model on the YCrCb color space to quickly learn the distribution of engineering images online, and design an engineering image recognizer based on multi-color space information.


2020 ◽  
Author(s):  
Soma Nonaka ◽  
Kei Majima ◽  
Shuntaro C. Aoki ◽  
Yukiyasu Kamitani

SummaryAchievement of human-level image recognition by deep neural networks (DNNs) has spurred interest in whether and how DNNs are brain-like. Both DNNs and the visual cortex perform hierarchical processing, and correspondence has been shown between hierarchical visual areas and DNN layers in representing visual features. Here, we propose the brain hierarchy (BH) score as a metric to quantify the degree of hierarchical correspondence based on the decoding of individual DNN unit activations from human brain activity. We find that BH scores for 29 pretrained DNNs with varying architectures are negatively correlated with image recognition performance, indicating that recently developed high-performance DNNs are not necessarily brain-like. Experimental manipulations of DNN models suggest that relatively simple feedforward architecture with broad spatial integration is critical to brain-like hierarchy. Our method provides new ways for designing DNNs and understanding the brain in consideration of their representational homology.


2019 ◽  
Vol 2019 ◽  
pp. 1-12 ◽  
Author(s):  
Tsun-Kuo Lin

This paper developed a principal component analysis (PCA)-integrated algorithm for feature identification in manufacturing; this algorithm is based on an adaptive PCA-based scheme for identifying image features in vision-based inspection. PCA is a commonly used statistical method for pattern recognition tasks, but an effective PCA-based approach for identifying suitable image features in manufacturing has yet to be developed. Unsuitable image features tend to yield poor results when used in conventional visual inspections. Furthermore, research has revealed that the use of unsuitable or redundant features might influence the performance of object detection. To address these problems, the adaptive PCA-based algorithm developed in this study entails the identification of suitable image features using a support vector machine (SVM) model for inspecting of various object images; this approach can be used for solving the inherent problem of detection that occurs when the extraction contains challenging image features in manufacturing processes. The results of experiments indicated that the proposed algorithm can successfully be used to adaptively select appropriate image features. The algorithm combines image feature extraction and PCA/SVM classification to detect patterns in manufacturing. The algorithm was determined to achieve high-performance detection and to outperform the existing methods.


2015 ◽  
Vol 2015 ◽  
pp. 1-7 ◽  
Author(s):  
Mohammad S. Khorsheed

Feature extraction plays an important role in text recognition as it aims to capture essential characteristics of the text image. Feature extraction algorithms widely range between robust and hard to extract features and noise sensitive and easy to extract features. Among those feature types are statistical features which are derived from the statistical distribution of the image pixels. This paper presents a novel method for feature extraction where simple statistical features are extracted from a one-pixel wide window that slides across the text line. The feature set is clustered in the feature space using vector quantization. The feature vector sequence is then injected to a classification engine for training and recognition purposes. The recognition system is applied to a data corpus which includes cursive Arabic text of more than 600 A4-size sheets typewritten in multiple computer-generated fonts. The system performance is compared to a previously published system from the literature with a similar engine but a different feature set.


2018 ◽  
Author(s):  
Clinton Haarlem ◽  
Rutger Vos

AbstractMosquitoes are important disease vectors. Different mosquito genera are associated with different diseases at varying levels of specificity. Hence, quick and low-cost methods of identification, even if relatively coarse and to genus level, will be of use in assessing risk and informing mitigation measures. Here we assess the extent to which digital photographs of mosquito wings taken with common cell phone cameras and clip-on lenses can be used to discriminate among mosquito genera when fed into image feature extraction algorithms. Our results show that genera may be distinguished on the basis of features extracted using the SURF algorithm. However, we also found that the naïve features examined here require very standardized photography and that different phone cameras have different signatures that may need to be taken into account.


Sign in / Sign up

Export Citation Format

Share Document