The Convolution Neural Network with Transformed Exponential Linear Unit Activation Function for Image Classification

Author(s):  
Edna C. Too ◽  
Yujian Li ◽  
Sam Njuki ◽  
Peter T. Yamak ◽  
Ting Zhang
Author(s):  
Jaya Gupta ◽  
◽  
Sunil Pathak ◽  
Gireesh Kumar

Image classification is critical and significant research problems in computer vision applications such as facial expression classification, satellite image classification, and plant classification based on images. Here in the paper, the image classification model is applied for identifying the display of daunting pictures on the internet. The proposed model uses Convolution neural network to identify these images and filter them through different blocks of the network, so that it can be classified accurately. The model will work as an extension to the web browser and will work on all websites when activated. The extension will be blurring the images and deactivating the links on web pages. This means that it will scan the entire web page and find all the daunting images present on that page. Then we will blur those images before they are loaded and the children could see them. Keywords— Activation Function, CNN, Images Classification , Optimizers, VGG-19


Author(s):  
Zenith Nandy

Abstract: In this paper, I built an AI model using deep learning, which identifies whether a given image is of an Arduino, a Beaglebone Black or a Jetson Nano. The identification of the object is based on prediction. The model is trained using 300 to 350 datasets of each category and is tested multiple times using different images at different angles, background colour and size. After multiple testing, the model is found to have 95 percent accuracy. Model used is Sequential and uses Convolution Neural Network (CNN) as its architecture. The activation function of each layer is RELU and for the output layer is Softmax. The output is a prediction and hence it is of probability type. This is a type of an application based project. The entire scripting is done using Python 3 programming language. Keywords: image classification, microcontroller boards, python, AI, deep learning, neural network


Author(s):  
T.K. Biryukova

Classic neural networks suppose trainable parameters to include just weights of neurons. This paper proposes parabolic integrodifferential splines (ID-splines), developed by author, as a new kind of activation function (AF) for neural networks, where ID-splines coefficients are also trainable parameters. Parameters of ID-spline AF together with weights of neurons are vary during the training in order to minimize the loss function thus reducing the training time and increasing the operation speed of the neural network. The newly developed algorithm enables software implementation of the ID-spline AF as a tool for neural networks construction, training and operation. It is proposed to use the same ID-spline AF for neurons in the same layer, but different for different layers. In this case, the parameters of the ID-spline AF for a particular layer change during the training process independently of the activation functions (AFs) of other network layers. In order to comply with the continuity condition for the derivative of the parabolic ID-spline on the interval (x x0, n) , its parameters fi (i= 0,...,n) should be calculated using the tridiagonal system of linear algebraic equations: To solve the system it is necessary to use two more equations arising from the boundary conditions for specific problems. For exam- ple the values of the grid function (if they are known) in the points (x x0, n) may be used for solving the system above: f f x0 = ( 0) , f f xn = ( n) . The parameters Iii+1 (i= 0,...,n−1 ) are used as trainable parameters of neural networks. The grid boundaries and spacing of the nodes of ID-spline AF are best chosen experimentally. The optimal selection of grid nodes allows improving the quality of results produced by the neural network. The formula for a parabolic ID-spline is such that the complexity of the calculations does not depend on whether the grid of nodes is uniform or non-uniform. An experimental comparison of the results of image classification from the popular FashionMNIST dataset by convolutional neural 0, x< 0 networks with the ID-spline AFs and the well-known ReLUx( ) =AF was carried out. The results reveal that the usage x x, ≥ 0 of the ID-spline AFs provides better accuracy of neural network operation than the ReLU AF. The training time for two convolutional layers network with two ID-spline AFs is just about 2 times longer than with two instances of ReLU AF. Doubling of the training time due to complexity of the ID-spline formula is the acceptable price for significantly better accuracy of the network. Wherein the difference of an operation speed of the networks with ID-spline and ReLU AFs will be negligible. The use of trainable ID-spline AFs makes it possible to simplify the architecture of neural networks without losing their efficiency. The modification of the well-known neural networks (ResNet etc.) by replacing traditional AFs with ID-spline AFs is a promising approach to increase the neural network operation accuracy. In a majority of cases, such a substitution does not require to train the network from scratch because it allows to use pre-trained on large datasets neuron weights supplied by standard software libraries for neural network construction thus substantially shortening training time.


IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 105659-105670 ◽  
Author(s):  
Rehan Ashraf ◽  
Muhammad Asif Habib ◽  
Muhammad Akram ◽  
Muhammad Ahsan Latif ◽  
Muhammad Sheraz Arshad Malik ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document