scholarly journals Separation of Commercially Important Tuna from Other Fishes Using Feature Descriptor and Pre-trained CNN Models

Author(s):  
Jisha Anu Jose ◽  
C. Sathish Kumar ◽  
S. Sureshkumar

Aims / Objectives: Identification of fish species is essential in export industries. Among the different fish species exported, tuna forms a significant portion and hence the separation of tuna from other fishes is necessary. The work aims to develop automated systems for the separation of commercially important tuna from other fishes.  Methodology: The work proposes two models for the classification of commercial fishes. The first model uses conventional feature descriptors, which extract features from both spatial and frequency domain. These features are combined and are reduced by an ensemble dimension reduction method. The combined and reduced feature sets are evaluated using different classifiers. The second proposed model uses four pre-trained convolutional neural networks, VGG16, VGG19, Xception, and MobileNet, for the classification. The models are fine-tuned for the classification process. Results: Results show that for the first model, extreme learning machine classifier with Mercer wavelet kernel gives high accuracy on combined feature set while the polynomial kernel ELM provides better performance with the reduced set. For the second model, a comparison of the performance of four CNN models is done, and results indicate that VGG19 outperforms other networks in the classification task.  Conclusion: Among the two proposed models, pre-trained CNN based model shows better performance than the conventional method in the separation task. Different performance measures, accuracy, precision, recall, F-score, and misclassification error are used to evaluate the system. A comparison of performance of the proposed models with the state-of-the-art systems is also reported.

Author(s):  
Jisha Anu Jose ◽  
C. Sathish Kumar

Automated recognition and classification of fishes are useful for studies dealing with counting of fishes for population assessments, discovering association between fishes and ecosystem, and monitoring of the ecosystem. This paper proposes a model which classifies the fishes belonging to the family Labridae in the genus and the species level. Features computed in the spatial and frequency domains are used in this work. All the images are preprocessed before feature extraction. Preprocessing step involves image segmentation for background elimination, de-noising and image enhancement. A combination of color, local binary pattern (LBP), histogram of oriented gradients (HOG), and wavelet features forms the feature vector. An ensemble feature reduction technique is used to reduce the attribute size. Performances of the system using combined as well as reduced feature sets are evaluated using seven popular classifiers. Among the classifiers, wavelet kernel extreme learning machine (ELM) showed higher classification accuracy of 96.65% in genus level and polynomial kernel ELM showed an accuracy of 92.42% in species level with the reduced feature set.


2021 ◽  
Vol 89 ◽  
pp. 22-32
Author(s):  
Awaz Ali ◽  
Dimitra Dodou ◽  
Gerwin Smit ◽  
Ruben Rink ◽  
Paul Breedveld

Author(s):  
Chaoqing Wang ◽  
Junlong Cheng ◽  
Yuefei Wang ◽  
Yurong Qian

A vehicle make and model recognition (VMMR) system is a common requirement in the field of intelligent transportation systems (ITS). However, it is a challenging task because of the subtle differences between vehicle categories. In this paper, we propose a hierarchical scheme for VMMR. Specifically, the scheme consists of (1) a feature extraction framework called weighted mask hierarchical bilinear pooling (WMHBP) based on hierarchical bilinear pooling (HBP) which weakens the influence of invalid background regions by generating a weighted mask while extracting features from discriminative regions to form a more robust feature descriptor; (2) a hierarchical loss function that can learn the appearance differences between vehicle brands, and enhance vehicle recognition accuracy; (3) collection of vehicle images from the Internet and classification of images with hierarchical labels to augment data for solving the problem of insufficient data and low picture resolution and improving the model’s generalization ability and robustness. We evaluate the proposed framework for accuracy and real-time performance and the experiment results indicate a recognition accuracy of 95.1% and an FPS (frames per second) of 107 for the framework for the Stanford Cars public dataset, which demonstrates the superiority of the method and its availability for ITS.


Sensors ◽  
2021 ◽  
Vol 21 (8) ◽  
pp. 2648
Author(s):  
Muhammad Aamir ◽  
Tariq Ali ◽  
Muhammad Irfan ◽  
Ahmad Shaf ◽  
Muhammad Zeeshan Azam ◽  
...  

Natural disasters not only disturb the human ecological system but also destroy the properties and critical infrastructures of human societies and even lead to permanent change in the ecosystem. Disaster can be caused by naturally occurring events such as earthquakes, cyclones, floods, and wildfires. Many deep learning techniques have been applied by various researchers to detect and classify natural disasters to overcome losses in ecosystems, but detection of natural disasters still faces issues due to the complex and imbalanced structures of images. To tackle this problem, we propose a multilayered deep convolutional neural network. The proposed model works in two blocks: Block-I convolutional neural network (B-I CNN), for detection and occurrence of disasters, and Block-II convolutional neural network (B-II CNN), for classification of natural disaster intensity types with different filters and parameters. The model is tested on 4428 natural images and performance is calculated and expressed as different statistical values: sensitivity (SE), 97.54%; specificity (SP), 98.22%; accuracy rate (AR), 99.92%; precision (PRE), 97.79%; and F1-score (F1), 97.97%. The overall accuracy for the whole model is 99.92%, which is competitive and comparable with state-of-the-art algorithms.


2021 ◽  
Vol 18 (1) ◽  
pp. 172988142098321
Author(s):  
Anzhu Miao ◽  
Feiping Liu

Human motion recognition is a branch of computer vision research and is widely used in fields like interactive entertainment. Most research work focuses on human motion recognition methods based on traditional video streams. Traditional RGB video contains rich colors, edges, and other information, but due to complex background, variable illumination, occlusion, viewing angle changes, and other factors, the accuracy of motion recognition algorithms is not high. For the problems, this article puts forward human motion recognition based on extreme learning machine (ELM). ELM uses the randomly calculated implicit network layer parameters for network training, which greatly reduces the time spent on network training and reduces computational complexity. In this article, the interframe difference method is used to detect the motion region, and then, the HOG3D feature descriptor is used for feature extraction. Finally, ELM is used for classification and recognition. The results imply that the method proposed here has achieved good results in human motion recognition.


2021 ◽  
Vol 11 (9) ◽  
pp. 3974
Author(s):  
Laila Bashmal ◽  
Yakoub Bazi ◽  
Mohamad Mahmoud Al Rahhal ◽  
Haikel Alhichri ◽  
Naif Al Ajlan

In this paper, we present an approach for the multi-label classification of remote sensing images based on data-efficient transformers. During the training phase, we generated a second view for each image from the training set using data augmentation. Then, both the image and its augmented version were reshaped into a sequence of flattened patches and then fed to the transformer encoder. The latter extracts a compact feature representation from each image with the help of a self-attention mechanism, which can handle the global dependencies between different regions of the high-resolution aerial image. On the top of the encoder, we mounted two classifiers, a token and a distiller classifier. During training, we minimized a global loss consisting of two terms, each corresponding to one of the two classifiers. In the test phase, we considered the average of the two classifiers as the final class labels. Experiments on two datasets acquired over the cities of Trento and Civezzano with a ground resolution of two-centimeter demonstrated the effectiveness of the proposed model.


Author(s):  
Jianfang Cao ◽  
Minmin Yan ◽  
Yiming Jia ◽  
Xiaodong Tian ◽  
Zibang Zhang

AbstractIt is difficult to identify the historical period in which some ancient murals were created because of damage due to artificial and/or natural factors; similarities in content, style, and color among murals; low image resolution; and other reasons. This study proposed a transfer learning-fused Inception-v3 model for dynasty-based classification. First, the model adopted Inception-v3 with frozen fully connected and softmax layers for pretraining over ImageNet. Second, the model fused Inception-v3 with transfer learning for parameter readjustment over small datasets. Third, the corresponding bottleneck files of the mural images were generated, and the deep-level features of the images were extracted. Fourth, the cross-entropy loss function was employed to calculate the loss value at each step of the training, and an algorithm for the adaptive learning rate on the stochastic gradient descent was applied to unify the learning rate. Finally, the updated softmax classifier was utilized for the dynasty-based classification of the images. On the constructed small datasets, the accuracy rate, recall rate, and F1 value of the proposed model were 88.4%, 88.36%, and 88.32%, respectively, which exhibited noticeable increases compared with those of typical deep learning models and modified convolutional neural networks. Comparisons of the classification outcomes for the mural dataset with those for other painting datasets and natural image datasets showed that the proposed model achieved stable classification outcomes with a powerful generalization capacity. The training time of the proposed model was only 0.7 s, and overfitting seldom occurred.


2002 ◽  
Vol 41 (04) ◽  
pp. 337-341 ◽  
Author(s):  
F. Cincotti ◽  
D. Mattia ◽  
C. Babiloni ◽  
F. Carducci ◽  
L. Bianchi ◽  
...  

Summary Objectives: In this paper, we explored the use of quadratic classifiers based on Mahalanobis distance to detect mental EEG patterns from a reduced set of scalp recording electrodes. Methods: Electrodes are placed in scalp centro-parietal zones (C3, P3, C4 and P4 positions of the international 10-20 system). A Mahalanobis distance classifier based on the use of full covariance matrix was used. Results: The quadratic classifier was able to detect EEG activity related to imagination of movement with an affordable accuracy (97% correct classification, on average) by using only C3 and C4 electrodes. Conclusions: Such a result is interesting for the use of Mahalanobis-based classifiers in the brain computer interface area.


2014 ◽  
Vol 643 ◽  
pp. 99-104
Author(s):  
Jin Yang ◽  
Yun Jie Li ◽  
Qin Li

In this paper, the process of the developments and changes of the network intrusion behaviors were analyzed. An improved epidemic spreading model was proposed to study the mechanisms of aggressive behaviors spreading, to predict the future course of an outbreak and to evaluate strategies to control a network epidemic. Based on Artificial Immune Systems, the concepts and formal definitions of immune cells were given. And in this paper, the forecasting algorithm based on Markov chain theory was proposed to improve the precision of network risk forecasting. The data of the Memory cells were analyzed directly and kinds of state-spaces were formed, which can be used to predict the risk of network situation by analyzing the cells status and the classification of optimal state. Experimental results show that the proposed model has the features of real-time processing for network situation awareness.


Sign in / Sign up

Export Citation Format

Share Document