scholarly journals Research on the Extraction of Image Edge Information in Convolutional Neural Networks

2021 ◽  
Vol 2083 (3) ◽  
pp. 032015
Author(s):  
Guanru Zou ◽  
Yulin Luo ◽  
Zefeng Feng

Abstract Convolutional neural network is an important neural network model in deep learning and a common algorithm in computer vision problems. From the perspective of practical application scenarios, this paper studies whether padding in convolutional neural network convolution layer weakens the image edge information. In order to eliminate the background factor, this paper select MNIST dataset as the research object, move the 0-9 digital image to the specified image edge by clearing the white area pixels in the specified direction, and use OpenCV to realize bilinear interpolation to scale the image to ensure that the image dimension is 28×28. The convolution neural network is built to train the original dataset and the processed dataset, and the accuracy rates are 0.9892 and 0.1082 respectively. In the comparative experiment, padding cannot solve the problem of weakening the image edge weight well. In the actual digital recognition scene, it is necessary to consider whether the core recognition area in the input image is at the edge of the image.

2020 ◽  
Vol 65 (6) ◽  
pp. 759-773
Author(s):  
Segu Praveena ◽  
Sohan Pal Singh

AbstractLeukaemia detection and diagnosis in advance is the trending topic in the medical applications for reducing the death toll of patients with acute lymphoblastic leukaemia (ALL). For the detection of ALL, it is essential to analyse the white blood cells (WBCs) for which the blood smear images are employed. This paper proposes a new technique for the segmentation and classification of the acute lymphoblastic leukaemia. The proposed method of automatic leukaemia detection is based on the Deep Convolutional Neural Network (Deep CNN) that is trained using an optimization algorithm, named Grey wolf-based Jaya Optimization Algorithm (GreyJOA), which is developed using the Grey Wolf Optimizer (GWO) and Jaya Optimization Algorithm (JOA) that improves the global convergence. Initially, the input image is applied to pre-processing and the segmentation is performed using the Sparse Fuzzy C-Means (Sparse FCM) clustering algorithm. Then, the features, such as Local Directional Patterns (LDP) and colour histogram-based features, are extracted from the segments of the pre-processed input image. Finally, the extracted features are applied to the Deep CNN for the classification. The experimentation evaluation of the method using the images of the ALL IDB2 database reveals that the proposed method acquired a maximal accuracy, sensitivity, and specificity of 0.9350, 0.9528, and 0.9389, respectively.


2020 ◽  
Vol 2020 ◽  
pp. 1-6
Author(s):  
Jian-ye Yuan ◽  
Xin-yuan Nan ◽  
Cheng-rong Li ◽  
Le-le Sun

Considering that the garbage classification is urgent, a 23-layer convolutional neural network (CNN) model is designed in this paper, with the emphasis on the real-time garbage classification, to solve the low accuracy of garbage classification and recycling and difficulty in manual recycling. Firstly, the depthwise separable convolution was used to reduce the Params of the model. Then, the attention mechanism was used to improve the accuracy of the garbage classification model. Finally, the model fine-tuning method was used to further improve the performance of the garbage classification model. Besides, we compared the model with classic image classification models including AlexNet, VGG16, and ResNet18 and lightweight classification models including MobileNetV2 and SuffleNetV2 and found that the model GAF_dense has a higher accuracy rate, fewer Params, and FLOPs. To further check the performance of the model, we tested the CIFAR-10 data set and found the accuracy rates of the model (GAF_dense) are 0.018 and 0.03 higher than ResNet18 and SufflenetV2, respectively. In the ImageNet data set, the accuracy rates of the model (GAF_dense) are 0.225 and 0.146 higher than Resnet18 and SufflenetV2, respectively. Therefore, the garbage classification model proposed in this paper is suitable for garbage classification and other classification tasks to protect the ecological environment, which can be applied to classification tasks such as environmental science, children’s education, and environmental protection.


Author(s):  
Ramesh Adhikari ◽  
Suresh Pokharel

Data augmentation is widely used in image processing and pattern recognition problems in order to increase the richness in diversity of available data. It is commonly used to improve the classification accuracy of images when the available datasets are limited. Deep learning approaches have demonstrated an immense breakthrough in medical diagnostics over the last decade. A significant amount of datasets are needed for the effective training of deep neural networks. The appropriate use of data augmentation techniques prevents the model from over-fitting and thus increases the generalization capability of the network while testing afterward on unseen data. However, it remains a huge challenge to obtain such a large dataset from rare diseases in the medical field. This study presents the synthetic data augmentation technique using Generative Adversarial Networks to evaluate the generalization capability of neural networks using existing data more effectively. In this research, the convolutional neural network (CNN) model is used to classify the X-ray images of the human chest in both normal and pneumonia conditions; then, the synthetic images of the X-ray from the available dataset are generated by using the deep convolutional generative adversarial network (DCGAN) model. Finally, the CNN model is trained again with the original dataset and augmented data generated using the DCGAN model. The classification performance of the CNN model is improved by 3.2% when the augmented data were used along with the originally available dataset.


2018 ◽  
Vol 7 (3.4) ◽  
pp. 127
Author(s):  
Nagaraj Bhat ◽  
U Eranna ◽  
Manoj Kumar Singh

In this paper, a complementary approach has applied to obtain the available edges in the image. The complementary image has obtained by subtracting the rough mirror mapped image from the input image. The universal approximation capability of feedforward neural network has applied to define the rough mirror mapping. Multilayer perceptron network and radial basis function network have considered obtaining the mapping. Effect of better learning has also explored in both network by applying adaptivenesss in their transform function available in the active nodes. Single image based training has given for few number of iterations in the development of mapping process.  It is observed that proposed method has self adjusted content aware oriented edge detection where as many existing methods like Sobel, Prewitt have shown their limitations in observing the edges associated with contents having similar shade in the surroundings.  


Diagnostics ◽  
2019 ◽  
Vol 9 (2) ◽  
pp. 38 ◽  
Author(s):  
Incheol Kim ◽  
Sivaramakrishnan Rajaraman ◽  
Sameer Antani

Deep learning (DL) methods are increasingly being applied for developing reliable computer-aided detection (CADe), diagnosis (CADx), and information retrieval algorithms. However, challenges in interpreting and explaining the learned behavior of the DL models hinders their adoption and use in real-world systems. In this study, we propose a novel method called “Class-selective Relevance Mapping” (CRM) for localizing and visualizing discriminative regions of interest (ROI) within a medical image. Such visualizations offer improved explanation of the convolutional neural network (CNN)-based DL model predictions. We demonstrate CRM effectiveness in classifying medical imaging modalities toward automatically labeling them for visual information retrieval applications. The CRM is based on linear sum of incremental mean squared errors (MSE) calculated at the output layer of the CNN model. It measures both positive and negative contributions of each spatial element in the feature maps produced from the last convolution layer leading to correct classification of an input image. A series of experiments on a “multi-modality” CNN model designed for classifying seven different types of image modalities shows that the proposed method is significantly better in detecting and localizing the discriminative ROIs than other state of the art class-activation methods. Further, to visualize its effectiveness we generate “class-specific” ROI maps by averaging the CRM scores of images in each modality class, and characterize the visual explanation through their different size, shape, and location for our multi-modality CNN model that achieved over 98% performance on a dataset constructed from publicly available images.


Sensors ◽  
2020 ◽  
Vol 20 (23) ◽  
pp. 6932
Author(s):  
Matthew Burns ◽  
Federico Cruciani ◽  
Philip Morrow ◽  
Chris Nugent ◽  
Sally McClean

The desire to remain living in one’s own home rather than a care home by those in need of 24/7 care is one that requires a level of understanding for the actions of an environment’s inhabitants. This can potentially be accomplished with the ability to recognise Activities of Daily Living (ADLs); however, this research focuses first on producing an unobtrusive solution for pose recognition where the preservation of privacy is a primary aim. With an accurate manner of predicting an inhabitant’s poses, their interactions with objects within the environment and, therefore, the activities they are performing, can begin to be understood. This research implements a Convolutional Neural Network (CNN), which has been designed with an original architecture derived from the popular AlexNet, to predict poses from thermal imagery that have been captured using thermopile infrared sensors (TISs). Five TISs have been deployed within the smart kitchen in Ulster University where each provides input to a corresponding trained CNN. The approach is evaluated using an original dataset and an F1-score of 0.9920 was achieved with all five TISs. The limitations of utilising a ceiling-based TIS are investigated and each possible permutation of corner-based TISs is evaluated to satisfy a trade-off between the number of TISs, the total sensor cost and the performances. These tests are also promising as F1-scores of 0.9266, 0.9149 and 0.8468 were achieved with the isolated use of four, three, and two corner TISs, respectively.


An Authenticated Security System is a highly desired feature. In this paper, a FreeHand Sketch-based Authentication Security strategy is proposed for authentication purposes by allowing a user to choose one label from a collection of different labels and asking him to sketch the corresponding image for the selected label for registration to avoid mischievous registration and the sketched image gets preprocessed using adaptive threshold with Gaussian mixture and then predicted with a trained Convolutional Neural Network(CNN) data model to generate the necessary image label. The produced image label will compare with selected image label. If both are same then the details will store in the system database. The user gets login with his/her authorized details with sketch based image password. The image password gets preprocessed using adaptive threshold with Gaussian mixture and then predicted with a trained CNN model to produce the image name. The produced image name will compare with the system database for authentication. The methodology is tested with some sample input image passwords and the performance calculation is carried out using metrics like Recall and Precision. The proposed work exhibits the accuracy of approximately 85% by ensuring the authentication for the user security.


2021 ◽  
Vol 7 ◽  
pp. e497
Author(s):  
Shakeel Shafiq ◽  
Tayyaba Azim

Deep neural networks have been widely explored and utilised as a useful tool for feature extraction in computer vision and machine learning. It is often observed that the last fully connected (FC) layers of convolutional neural network possess higher discrimination power as compared to the convolutional and maxpooling layers whose goal is to preserve local and low-level information of the input image and down sample it to avoid overfitting. Inspired from the functionality of local binary pattern (LBP) operator, this paper proposes to induce discrimination into the mid layers of convolutional neural network by introducing a discriminatively boosted alternative to pooling (DBAP) layer that has shown to serve as a favourable replacement of early maxpooling layer in a convolutional neural network (CNN). A thorough research of the related works show that the proposed change in the neural architecture is novel and has not been proposed before to bring enhanced discrimination and feature visualisation power achieved from the mid layer features. The empirical results reveal that the introduction of DBAP layer in popular neural architectures such as AlexNet and LeNet produces competitive classification results in comparison to their baseline models as well as other ultra-deep models on several benchmark data sets. In addition, better visualisation of intermediate features can allow one to seek understanding and interpretation of black box behaviour of convolutional neural networks, used widely by the research community.


2021 ◽  
Vol 8 (3) ◽  
pp. 533
Author(s):  
Budi Nugroho ◽  
Eva Yulia Puspaningrum

<p class="Abstrak">Saat ini banyak dikembangkan proses pendeteksian pneumonia berdasarkan citra paru-paru dari hasil foto rontgen (x-ray), sebagaimana juga dilakukan pada penelitian ini. Metode yang digunakan adalah <em>Convolutional Neural Network</em> (CNN) dengan arsitektur yang berbeda dengan sejumlah penelitian sebelumnya. Selain itu, penelitian ini juga memodifikasi model CNN dimana metode <em>Extreme Learning Machine</em> (ELM) digunakan pada bagian klasifikasi, yang kemudian disebut CNN-ELM. Dataset untuk uji coba menggunakan kumpulan citra paru-paru hasil foto rontgen pada Kaggle yang terdiri atas 1.583 citra normal dan 4.237 citra pneumonia. Citra asal pada dataset kaggle ini bervariasi, tetapi hampir semua diatas ukuran 1000x1000 piksel. Ukuran citra yang besar ini dapat membuat pemrosesan klasifikasi kurang efektif, sehingga mesin CNN biasanya memodifikasi ukuran citra menjadi lebih kecil. Pada penelitian ini, pengujian dilakukan dengan variasi ukuran citra input, untuk mengetahui pengaruhnya terhadap kinerja mesin pengklasifikasi. Hasil uji coba menunjukkan bahwa ukuran citra input berpengaruh besar terhadap kinerja klasifikasi pneumonia, baik klasifikasi yang menggunakan metode CNN maupun CNN-ELM. Pada ukuran citra input 200x200, metode CNN dan CNN-ELM menunjukkan kinerja paling tinggi. Jika kinerja kedua metode itu dibandingkan, maka Metode CNN-ELM menunjukkan kinerja yang lebih baik daripada CNN pada semua skenario uji coba. Pada kondisi kinerja paling tinggi, selisih akurasi antara metode CNN-ELM dan CNN mencapai 8,81% dan selisih F1 Score mencapai 0,0729. Hasil penelitian ini memberikan informasi penting bahwa ukuran citra input memiliki pengaruh besar terhadap kinerja klasifikasi pneumonia, baik klasifikasi menggunakan metode CNN maupun CNN-ELM. Selain itu, pada semua ukuran citra input yang digunakan untuk proses klasifikasi, metode CNN-ELM menunjukkan kinerja yang lebih baik daripada metode CNN.</p><p class="Abstrak"> </p><p class="Abstrak"><em><strong>Abstract</strong></em></p><p class="Abstract"><em>This research developed a pneumonia detection machine based on the lungs' images from X-rays (x-rays). The method used is the Convolutional Neural Network (CNN) with a different architecture from some previous research. Also, the CNN model is modified, where the classification process uses the Extreme Learning Machine (ELM), which is then called the CNN-ELM method. The empirical experiments dataset used a collection of lung x-ray images on Kaggle consisting of 1,583 normal images and 4,237 pneumonia images. The original image's size on the Kaggle dataset varies, but almost all of the images are more than 1000x1000 pixels. For classification processing to be more effective, CNN machines usually use reduced-size images. In this research, experiments were carried out with various input image sizes to determine the effect on the classifier's performance. The experimental results show that the input images' size has a significant effect on the classification performance of pneumonia, both the CNN and CNN-ELM classification methods. At the 200x200 input image size, the CNN and CNN-ELM methods showed the highest performance. If the two methods' performance is compared, then the CNN-ELM Method shows better performance than CNN in all test scenarios. The difference in accuracy between the CNN-ELM and CNN methods reaches 8.81% at the highest performance conditions, and the difference in F1-Score reaches 0.0729. This research provides important information that the size of the input image has a major influence on the classification performance of pneumonia, both classification using the CNN and CNN-ELM methods. Also, on all input image sizes used for the classification process, the CNN-ELM method shows better performance than the CNN method.</em></p>


Sign in / Sign up

Export Citation Format

Share Document