scholarly journals Optimizing Convolutional Neural Network Parameters for Better Image Classification

Author(s):  
Manik Dhingra ◽  
Sarthak Rawat ◽  
Jinan Fiaidhi

The work presented here works on getting higher performances for image recognition task using convolutional neural networks on the MNIST handwritten digits data-set. A range of techniques are compared for improvements with respect to time and accuracy, such as using one-shot Extreme Learning Machines (ELM) in place of the iteratively tuned fully-connected networks for classification, using transfer learning for faster convergence of image classification, and improving the size of data-set and making robust models by image augmentation. The final implementation is hosted on cloud as a web-service for better visualization of the prediction results.

2020 ◽  
Author(s):  
Manik Dhingra ◽  
Sarthak Rawat ◽  
Jinan Fiaidhi

The work presented here works on getting higher performances for image recognition task using convolutional neural networks on the MNIST handwritten digits data-set. A range of techniques are compared for improvements with respect to time and accuracy, such as using one-shot Extreme Learning Machines (ELM) in place of the iteratively tuned fully-connected networks for classification, using transfer learning for faster convergence of image classification, and improving the size of data-set and making robust models by image augmentation. The final implementation is hosted on cloud as a web-service for better visualization of the prediction results.


Inventions ◽  
2021 ◽  
Vol 6 (4) ◽  
pp. 70
Author(s):  
Elena Solovyeva ◽  
Ali Abdullah

In this paper, the structure of a separable convolutional neural network that consists of an embedding layer, separable convolutional layers, convolutional layer and global average pooling is represented for binary and multiclass text classifications. The advantage of the proposed structure is the absence of multiple fully connected layers, which is used to increase the classification accuracy but raises the computational cost. The combination of low-cost separable convolutional layers and a convolutional layer is proposed to gain high accuracy and, simultaneously, to reduce the complexity of neural classifiers. Advantages are demonstrated at binary and multiclass classifications of written texts by means of the proposed networks under the sigmoid and Softmax activation functions in convolutional layer. At binary and multiclass classifications, the accuracy obtained by separable convolutional neural networks is higher in comparison with some investigated types of recurrent neural networks and fully connected networks.


2021 ◽  
Vol 2021 ◽  
pp. 1-11
Author(s):  
Haibin Chang ◽  
Ying Cui

More and more image materials are used in various industries these days. Therefore, how to collect useful images from a large set has become an urgent priority. Convolutional neural networks (CNN) have achieved good results in certain image classification tasks, but there are still problems such as poor classification ability, low accuracy, and slow convergence speed. This article mainly introduces the image classification algorithm (ICA) research based on the multilabel learning of the improved convolutional neural network and some improvement ideas for the research of the ICA based on the multilabel learning of the convolutional neural network. This paper proposes an ICA research method based on multilabel learning of improved convolutional neural networks, including the image classification process, convolutional network algorithm, and multilabel learning algorithm. The conclusions show that the average maximum classification accuracy of the improved CNN in this paper is 90.63%, and the performance is better, which is beneficial to improving the efficiency of image classification. The improved CNN network structure has reached the highest accuracy rate of 91.47% on the CIFAR-10 data set, which is much higher than the traditional CNN algorithm.


2020 ◽  
Vol 2 (2) ◽  
pp. 23
Author(s):  
Lei Wang

<p>As an important research achievement in the field of brain like computing, deep convolution neural network has been widely used in many fields such as computer vision, natural language processing, information retrieval, speech recognition, semantic understanding and so on. It has set off a wave of neural network research in industry and academia and promoted the development of artificial intelligence. At present, the deep convolution neural network mainly simulates the complex hierarchical cognitive laws of the human brain by increasing the number of layers of the network, using a larger training data set, and improving the network structure or training learning algorithm of the existing neural network, so as to narrow the gap with the visual system of the human brain and enable the machine to acquire the capability of "abstract concepts". Deep convolution neural network has achieved great success in many computer vision tasks such as image classification, target detection, face recognition, pedestrian recognition, etc. Firstly, this paper reviews the development history of convolutional neural networks. Then, the working principle of the deep convolution neural network is analyzed in detail. Then, this paper mainly introduces the representative achievements of convolution neural network from the following two aspects, and shows the improvement effect of various technical methods on image classification accuracy through examples. From the aspect of adding network layers, the structures of classical convolutional neural networks such as AlexNet, ZF-Net, VGG, GoogLeNet and ResNet are discussed and analyzed. From the aspect of increasing the size of data set, the difficulties of manually adding labeled samples and the effect of using data amplification technology on improving the performance of neural network are introduced. This paper focuses on the latest research progress of convolution neural network in image classification and face recognition. Finally, the problems and challenges to be solved in future brain-like intelligence research based on deep convolution neural network are proposed.</p>


2021 ◽  
Vol 2127 (1) ◽  
pp. 012024
Author(s):  
T E Razumov ◽  
D V Churikov ◽  
O V Kravchenko

Abstract In this paper, the problem of constructing a model for detecting and filtering unwanted spam messages is solved. A fully connected convolutional neural network (FCNN) was chosen as the model of the classifier of unwanted emails in email. It allows you to divide emails into two categories: spam and not spam. The main result of the research is a software application in the C++ language, which has a micro-service architecture and solves the problem of image classification. The app can handle more than 106 requests per minute in real-time.


2020 ◽  
Vol 86 (20) ◽  
Author(s):  
Kaidi Wang ◽  
Lei Chen ◽  
Xiangyun Ma ◽  
Lina Ma ◽  
Keng C. Chou ◽  
...  

ABSTRACT Rapid and accurate identification of Arcobacter is of great importance because it is considered an emerging food- and waterborne pathogen and potential zoonotic agent. Raman spectroscopy can differentiate bacteria based on Raman scattering spectral patterns of whole cells in a fast, reagentless, and easy-to-use manner. We aimed to detect and discriminate Arcobacter bacteria at the species level using confocal micro-Raman spectroscopy (785 nm) coupled with neural networks. A total of 82 reference and field isolates of 18 Arcobacter species from clinical, environmental, and agri-food sources were included. We determined that the bacterial cultivation time and growth temperature did not significantly influence the Raman spectral reproducibility and discrimination capability. The genus Arcobacter could be successfully differentiated from the closely related genera Campylobacter and Helicobacter using principal-component analysis. For the identification of Arcobacter to the species level, an accuracy of 97.2% was achieved for all 18 Arcobacter species using Raman spectroscopy combined with a convolutional neural network (CNN). The predictive capability of Raman-CNN was further validated using an independent data set of 12 Arcobacter strains. Furthermore, a Raman spectroscopy-based fully connected artificial neural network (ANN) was constructed to determine the actual ratio of a specific Arcobacter species in a bacterial mixture ranging from 5% to 100% by biomass (regression coefficient >0.99). The application of both CNN and fully connected ANN improved the accuracy of Raman spectroscopy for bacterial species determination compared to the conventional chemometrics. This newly developed approach enables rapid identification and species determination of Arcobacter within an hour following cultivation. IMPORTANCE Rapid identification of bacterial pathogens is critical for developing an early warning system and performing epidemiological investigation. Arcobacter is an emerging foodborne pathogen and has become more important in recent decades. The incidence of Arcobacter species in the agro-ecosystem is probably underestimated mainly due to the limitation in the available detection and characterization techniques. Raman spectroscopy combined with machine learning can accurately identify Arcobacter at the species level in a rapid and reliable manner, providing a promising tool for epidemiological surveillance of this microbe in the agri-food chain. The knowledge elicited from this study has the potential to be used for routine bacterial screening and diagnostics by the government, food industry, and clinics.


Author(s):  
Panagiotis Kouvaros ◽  
Alessio Lomuscio

We introduce an efficient method for the complete verification of ReLU-based feed-forward neural networks. The method implements branching on the ReLU states on the basis of a notion of dependency between the nodes. This results in dividing the original verification problem into a set of sub-problems whose MILP formulations require fewer integrality constraints. We evaluate the method on all of the ReLU-based fully connected networks from the first competition for neural network verification. The experimental results obtained show 145% performance gains over the present state-of-the-art in complete verification.


1994 ◽  
Vol 37 (3) ◽  
Author(s):  
G. Romeo

Pattern recognition belongs to a class of Problems which are easily solved by humans, but difficult for computers. It is sometimes difficult to formalize a problem which a human operator can casily understand by using examples. Neural networks are useful in solving this kind of problem. A neural network may, under certain conditions, simulate a well trained human operator in recognizing different types of earthquakes or in detecting the presence of a seismic event. It is then shown how a fully connected multi layer perceptron may perform a recognition task. It is shown how a self training auto associative neural network may detect an earthquake occurrence analysing the change in signal characteristics.


Vestnik MEI ◽  
2021 ◽  
Vol 3 (3) ◽  
pp. 103-109
Author(s):  
Andrey I. Mamontov ◽  

In solving the classification problem, a fully connected trainable neural network (with adjusting the parameters represented by double-precision real numbers) is used as a mathematical model. After the training is completed, the neural network parameters are rounded and represented as fixed-point numbers (integers). The aim of the study is to reduce the required amount of the computing system memory for storing the obtained integer parameters. To reduce the amount of memory, the following methods for storing integer parameters are developed, which are based on representing the linear polynomials included in a fully connected neural network using compositions of simpler functions: - a method based on representing the considered polynomial as a sum of simpler polynomials; - a method based on separately storing the information about additions and multiplications. In the experiment with the MNIST data set, it took 1.41 MB to store real parameters of a fully connected neural network, 0.7 MB to store integer parameters without using the proposed methods, 0.47 MB in the RAM and 0.3 MB in compressed form on the disk when using the first method, and 0.25 MB on the disk when using the second method. In the experiment with the USPS data set, it took 0.25 MB to store real parameters of a fully connected neural network, 0.1 MB to store integer parameters without using the proposed methods, 0.05 MB in the RAM and approximately the same amount in compressed form on the disk when using the first method, and 0.03 MB on the disk when using the second method. The study results can be applied in using fully connected neural networks to solve various recognition problems under the conditions of limited hardware capacities.


Sign in / Sign up

Export Citation Format

Share Document