scholarly journals Automated Individual Cattle Identification Using Video Data: A Unified Deep Learning Architecture Approach

2021 ◽  
Vol 2 ◽  
Author(s):  
Yongliang Qiao ◽  
Cameron Clark ◽  
Sabrina Lomax ◽  
He Kong ◽  
Daobilige Su ◽  
...  

Individual cattle identification is a prerequisite and foundation for precision livestock farming. Existing methods for cattle identification require radio frequency or visual ear tags, all of which are prone to loss or damage. Here, we propose and implement a new unified deep learning approach to cattle identification using video analysis. The proposed deep learning framework is composed of a Convolutional Neural Network (CNN) and Bidirectional Long Short-Term Memory (BiLSTM) with a self-attention mechanism. More specifically, the Inception-V3 CNN was used to extract features from a cattle video dataset taken in a feedlot with rear-view. Extracted features were then fed to a BiLSTM layer to capture spatio-temporal information. Then, self-attention was employed to provide a different focus on the features captured by BiLSTM for the final step of cattle identification. We used a total of 363 rear-view videos from 50 cattle at three different times with an interval of 1 month between data collection periods. The proposed method achieved 93.3% identification accuracy using a 30-frame video length, which outperformed current state-of-the-art methods (Inception-V3, MLP, SimpleRNN, LSTM, and BiLSTM). Furthermore, two different attention schemes, namely, additive and multiplicative attention mechanisms were compared. Our results show that the additive attention mechanism achieved 93.3% accuracy and 91.0% recall, greater than multiplicative attention mechanism with 90.7% accuracy and 87.0% recall. Video length also impacted accuracy, with video sequence length up to 30-frames enhancing identification performance. Overall, our approach can capture key spatio-temporal features to improve cattle identification accuracy, enabling automated cattle identification for precision livestock farming.

2021 ◽  
Vol 64 (6) ◽  
pp. 1823-1833
Author(s):  
Yangyang Guo ◽  
Yongliang Qiao ◽  
Salah Sukkarieh ◽  
Lilong Chai ◽  
Dongjian He

HighlightsBiGRU-attention based cow behavior classification was proposed.Key spatial-temporal features were captured for behavior representation.BiGRU-attention achieved >82% classification accuracy on calf and adult cow datasets.The proposed method could be used for similar animal behavior classification.Abstract. Animal behavior consists of time series activities, which can reflect animals’ health and welfare status. Monitoring and classifying animal behavior facilitates management decisions to optimize animal performance, welfare, and environmental outcomes. In recent years, deep learning methods have been applied to monitor animal behavior worldwide. To achieve high behavior classification accuracy, a BiGRU-attention based method is proposed in this article to classify some common behaviors, such as exploring, feeding, grooming, standing, and walking. In our work, (1) Inception-V3 was first applied to extract convolutional neural network (CNN) features for each image frame in videos, (2) bidirectional gated recurrent unit (BiGRU) was used to further extract spatial-temporal features, (3) an attention mechanism was deployed to allocate weights to each of the extracted spatial-temporal features according to feature similarity, and (4) the weighted spatial-temporal features were fed to a Softmax layer for behavior classification. Experiments were conducted on two datasets (i.e., calf and adult cow), and the proposed method achieved 82.35% and 82.26% classification accuracy on the calf and adult cow datasets, respectively. In addition, in comparison with other methods, the proposed BiGRU-attention method outperformed long short-term memory (LSTM), bidirectional LSTM (BiLSTM), and BiGRU. Overall, the proposed BiGRU-attention method can capture key spatial-temporal features to significantly improve animal behavior classification, which is favorable for automatic behavior classification in precision livestock farming. Keywords: BiGRU, Cow behavior, Deep learning, LSTM, Precision livestock farming.


2021 ◽  
Author(s):  
Yu Rang Park ◽  
Sang Ho Hwang ◽  
Yeonsoo Yu ◽  
Jichul Kim ◽  
Taeyeop Lee ◽  
...  

BACKGROUND Early detection and intervention of developmental disabilities (DDs) are critical for improving the long-term outcomes of the afflicted children. Mobile-based applications are easily accessible and may thus help the early identification of DDs. OBJECTIVE We aimed to identify facial expression and head pose based on face landmark data extracted from face recording videos and to differentiate the characteristics between children with DDs and those without. METHODS Eighty-nine children (DD, n=33; typically developing, n=56) were included in the analysis. Using the mobile-based application, we extracted facial landmarks and head poses from the recorded videos and performed Long Short-Term Memory(LSTM)-based DD classification. RESULTS Stratified k-fold cross-validation showed that the average values of accuracy, precision, recall, and f1-score of the LSTM based deep learning model of DD children were 88%, 91%,72%, and 80%, respectively. Through the interpretation of prediction results using SHapley Additive exPlanations (SHAP), we confirmed that the nodding head angle variable was the most important variable. All of the top 10 variables of importance had significant differences in the distribution between children with DDs and those without (p<0.05). CONCLUSIONS Our results provide preliminary evidence that the deep-learning classification model using mobile-based children’s video data could be used for the early detection of children with DDs.


2021 ◽  
Vol 11 (21) ◽  
pp. 10249
Author(s):  
Chien-Nguyen Nhu ◽  
Minho Park

Cloud computing is currently considered the most cost-effective platform for offering business and consumer IT services over the Internet. However, it is prone to new vulnerabilities. A new type of attack called an economic denial of sustainability (EDoS) attack exploits the pay-per-use model to scale up the resource usage over time to the extent that the cloud user has to pay for the unexpected usage charge. To prevent EDoS attacks, a few solutions have been proposed, including hard-threshold and machine learning-based solutions. Among them, long short-term memory (LSTM)-based solutions achieve much higher accuracy and false-alarm rates than hard-threshold and other machine learning-based solutions. However, LSTM requires a long sequence length of the input data, leading to a degraded performance owing to increases in the calculations, the detection time, and consuming a large number of computing resources of the defense system. We, therefore, propose a two-phase deep learning-based EDoS detection scheme that uses an LSTM model to detect each abnormal flow in network traffic; however, the LSTM model requires only a short sequence length of five of the input data. Thus, the proposed scheme can take advantage of the efficiency of the LSTM algorithm in detecting each abnormal flow in network traffic, while reducing the required sequence length of the input data. A comprehensive performance evaluation shows that our proposed scheme outperforms the existing solutions in terms of accuracy and resource consumption.


Author(s):  
Yuqi Yu ◽  
Hanbing Yan ◽  
Yuan Ma ◽  
Hao Zhou ◽  
Hongchao Guan

AbstractHypertext Transfer Protocol (HTTP) accounts for a large portion of Internet application-layer traffic. Since the payload of HTTP traffic can record website status and user request information, many studies use HTTP protocol traffic for web application attack detection. In this work, we propose DeepHTTP, an HTTP traffic detection framework based on deep learning. Unlike previous studies, this framework not only performs malicious traffic detection but also uses the deep learning model to mine malicious fields of the traffic payload. The detection model is called AT-Bi-LSTM, which is based on Bidirectional Long Short-Term Memory (Bi-LSTM) with attention mechanism. The attention mechanism can improve the discriminative ability and make the result interpretable. To enhance the generalization ability of the model, this paper proposes a novel feature extraction method. Experiments show that DeepHTTP has an excellent performance in malicious traffic discrimination and pattern mining.


2020 ◽  
Vol 2020 ◽  
pp. 1-12
Author(s):  
Zhenbo Lu ◽  
Wei Zhou ◽  
Shixiang Zhang ◽  
Chen Wang

Quick and accurate crash detection is important for saving lives and improved traffic incident management. In this paper, a feature fusion-based deep learning framework was developed for video-based urban traffic crash detection task, aiming at achieving a balance between detection speed and accuracy with limited computing resource. In this framework, a residual neural network (ResNet) combined with attention modules was proposed to extract crash-related appearance features from urban traffic videos (i.e., a crash appearance feature extractor), which were further fed to a spatiotemporal feature fusion model, Conv-LSTM (Convolutional Long Short-Term Memory), to simultaneously capture appearance (static) and motion (dynamic) crash features. The proposed model was trained by a set of video clips covering 330 crash and 342 noncrash events. In general, the proposed model achieved an accuracy of 87.78% on the testing dataset and an acceptable detection speed (FPS > 30 with GTX 1060). Thanks to the attention module, the proposed model can capture the localized appearance features (e.g., vehicle damage and pedestrian fallen-off) of crashes better than conventional convolutional neural networks. The Conv-LSTM module outperformed conventional LSTM in terms of capturing motion features of crashes, such as the roadway congestion and pedestrians gathering after crashes. Compared to traditional motion-based crash detection model, the proposed model achieved higher detection accuracy. Moreover, it could detect crashes much faster than other feature fusion-based models (e.g., C3D). The results show that the proposed model is a promising video-based urban traffic crash detection algorithm that could be used in practice in the future.


2020 ◽  
Vol 81 ◽  
pp. 149-165
Author(s):  
H Apaydin ◽  
MT Sattari

It is clearly known that precipitation is essential for fauna and flora. Studies have shown that location and temporal factors have an effect on precipitation. Accurate prediction of precipitation is very important for water resource management, and artificial intelligence methods are frequently used to make such predictions. In this study, the deep-learning and geographic information system (GIS) hybrid approach based on spatio-temporal variables was applied in order to model the amount of precipitation on Turkey’s coastline. Information about latitude, longitude, altitude, distance to the sea, and aspect was taken from meteorological stations, and these factors were utilized as spatial variables. The change in monthly precipitation was taken into account as a temporal variable. Artificial intelligence methods such as Gaussian process regression, support vector regression, the Broyden-Fletcher-Goldfarb-Shanno artificial neural network, M5, random forest, and long short-term memory (LSTM) were used. According to the results of the study, in which different input variable alternatives were also evaluated, LSTM was the most successful method for predicting precipitation with a value of 0.93 R. The study shows that the amount of precipitation can be estimated and a distribution map can be drawn by using spatio-temporal data and the deep-learning and GIS hybrid method at points where the measurement is not performed.


Sign in / Sign up

Export Citation Format

Share Document