Deep Learning for Printed Mottle Defect Grading

2020 ◽  
Vol 2020 (8) ◽  
pp. 184-1-184-9
Author(s):  
Jianhang Chen ◽  
Qian Lin ◽  
Jan P. Allebach

In this paper, we propose a new method for printed mottle defect grading. By training the data scanned from printed images, our deep learning method based on a Convolutional Neural Network (CNN) can classify various images with different mottle defect levels. Different from traditional methods to extract the image features, our method utilizes a CNN for the first time to extract the features automatically without manual feature design. Different data augmentation methods such as rotation, flip, zoom, and shift are also applied to the original dataset. The final network is trained by transfer learning using the ResNet-34 network pretrained on the ImageNet dataset connected with fully connected layers. The experimental results show that our approach leads to a 13.16% error rate in the T dataset, which is a dataset with a single image content, and a 20.73% error rate in a combined dataset with different contents.

Author(s):  
Uzma Batool ◽  
Mohd Ibrahim Shapiai ◽  
Nordinah Ismail ◽  
Hilman Fauzi ◽  
Syahrizal Salleh

Silicon wafer defect data collected from fabrication facilities is intrinsically imbalanced because of the variable frequencies of defect types. Frequently occurring types will have more influence on the classification predictions if a model gets trained on such skewed data. A fair classifier for such imbalanced data requires a mechanism to deal with type imbalance in order to avoid biased results. This study has proposed a convolutional neural network for wafer map defect classification, employing oversampling as an imbalance addressing technique. To have an equal participation of all classes in the classifier’s training, data augmentation has been employed, generating more samples in minor classes. The proposed deep learning method has been evaluated on a real wafer map defect dataset and its classification results on the test set returned a 97.91% accuracy. The results were compared with another deep learning based auto-encoder model demonstrating the proposed method, a potential approach for silicon wafer defect classification that needs to be investigated further for its robustness.


2021 ◽  
pp. 1-11
Author(s):  
Sunil Rao ◽  
Vivek Narayanaswamy ◽  
Michael Esposito ◽  
Jayaraman J. Thiagarajan ◽  
Andreas Spanias

Reliable and rapid non-invasive testing has become essential for COVID-19 diagnosis and tracking statistics. Recent studies motivate the use of modern machine learning (ML) and deep learning (DL) tools that utilize features of coughing sounds for COVID-19 diagnosis. In this paper, we describe system designs that we developed for COVID-19 cough detection with the long-term objective of embedding them in a testing device. More specifically, we use log-mel spectrogram features extracted from the coughing audio signal and design a series of customized deep learning algorithms to develop fast and automated diagnosis tools for COVID-19 detection. We first explore the use of a deep neural network with fully connected layers. Additionally, we investigate prospects of efficient implementation by examining the impact on the detection performance by pruning the fully connected neural network based on the Lottery Ticket Hypothesis (LTH) optimization process. In general, pruned neural networks have been shown to provide similar performance gains to that of unpruned networks with reduced computational complexity in a variety of signal processing applications. Finally, we investigate the use of convolutional neural network architectures and in particular the VGG-13 architecture which we tune specifically for this application. Our results show that a unique ensembling of the VGG-13 architecture trained using a combination of binary cross entropy and focal losses with data augmentation significantly outperforms the fully connected networks and other recently proposed baselines on the DiCOVA 2021 COVID-19 cough audio dataset. Our customized VGG-13 model achieves an average validation AUROC of 82.23% and a test AUROC of 78.3% at a sensitivity of 80.49%.


2021 ◽  
Author(s):  
Danial Sharifrazi ◽  
Roohallah Alizadehsani ◽  
Navid Hoseini Izadi ◽  
Mohamad Roshanzamir ◽  
Afshin Shoeibi ◽  
...  

Abstract Hypertrophic cardiomyopathy (HCM) can lead to serious cardiac problems. HCM is often diagnosed by an expert using cardiovascular magnetic resonance (CMR) images obtained from patients. In this research, we aimed to develop a deep learning technique to automate HCM diagnosis. CMR images of 37421 healthy and 21846 HCM patients were obtained during two years. Images obtained from female patients form 53% of the collected dataset. The mean and standard deviation of the dataset patients’ age are 48.2 and 19.5 years, respectively. Three experts inspected images and determined whether a case has HCM or not. New data augmentation was used to generate new images by employing color filtering on the existing ones. To classify the augmented images, we used a deep convolutional neural network (CNN). To the best of our knowledge, this is the first time CNN is used for HCM diagnosis. We designed our CNN from scratch to reach acceptable diagnosis accuracy. Comparing the designed algorithm output with the experts’ opinions, the method could achieve accuracy of 95.23%, recall of 97.90%, and specificity of 93.06% on the original dataset. The same performance metrics for the designed algorithm on the augmented dataset were 98.53%, 98.70%, and 95.21%, respectively. We have also experimented with different optimizers (e.g. Adadelta and Adagrad) and other data augmentation methods (e.g. height shift and rotation) to further evaluate the proposed method. Using our data augmentation method, accuracy of 98.53% were achieved which is higher than the best accuracy (95.83%) obtained by the other data augmentation methods which have been evaluated. The upper bound on difference between true error rate and empirical error rate of the proposed method has also been provided in order to present better performance analysis. The advantages of employing the proposed method are elimination of contrast agent and its complications, decreased CMR examination time, lower costs for patients and cardiac imaging centers.


2020 ◽  
Vol 8 (11) ◽  
pp. 924
Author(s):  
Guan Wei Thum ◽  
Sai Hong Tang ◽  
Siti Azfanizam Ahmad ◽  
Moath Alrifaey

Underwater cables or pipelines are commonly utilized elements in ocean research, marine engineering, power transmission, and communication-based activities. Their performance necessitates regularly conducted inspection for maintenance purposes. A vision system is commonly used by autonomous underwater vehicles (AUVs) to track and search for underwater cable. Its traditional methods are characteristically applicable in AUVs, wherein they are equipped with handcrafted features and shallow trainable architectures. However, such methods are subpar or even incapable of tracking underwater cable in fast-changing and complex underwater conditions. In contrast to this, the deep learning method is linked with the capacity to learn semantic, high-level, and deeper features, thus rendering it recommended for performing underwater cable tracking. In this study, several deep Convolutional Neural Network (CNN) models were proposed to classify underwater cable images obtained from a set of underwater images, whereby transfer learning and data augmentation were applied to enhance the classification accuracy. Following a comparison and discussion regarding the performance of these models, MobileNetV2 outperformed among other models and yielded lower computational time and the highest accuracy for classifying underwater cable images at 93.5%. Hence, the main contribution of this study is geared toward developing a deep learning method for underwater cable image classification.


2021 ◽  
pp. 1-11
Author(s):  
Yaning Liu ◽  
Lin Han ◽  
Hexiang Wang ◽  
Bo Yin

Papillary thyroid carcinoma (PTC) is a common carcinoma in thyroid. As many benign thyroid nodules have the papillary structure which could easily be confused with PTC in morphology. Thus, pathologists have to take a lot of time on differential diagnosis of PTC besides personal diagnostic experience and there is no doubt that it is subjective and difficult to obtain consistency among observers. To address this issue, we applied deep learning to the differential diagnosis of PTC and proposed a histological image classification method for PTC based on the Inception Residual convolutional neural network (IRCNN) and support vector machine (SVM). First, in order to expand the dataset and solve the problem of histological image color inconsistency, a pre-processing module was constructed that included color transfer and mirror transform. Then, to alleviate overfitting of the deep learning model, we optimized the convolution neural network by combining Inception Network and Residual Network to extract image features. Finally, the SVM was trained via image features extracted by IRCNN to perform the classification task. Experimental results show effectiveness of the proposed method in the classification of PTC histological images.


2021 ◽  
Vol 11 (12) ◽  
pp. 5488
Author(s):  
Wei Ping Hsia ◽  
Siu Lun Tse ◽  
Chia Jen Chang ◽  
Yu Len Huang

The purpose of this article is to evaluate the accuracy of the optical coherence tomography (OCT) measurement of choroidal thickness in healthy eyes using a deep-learning method with the Mask R-CNN model. Thirty EDI-OCT of thirty patients were enrolled. A mask region-based convolutional neural network (Mask R-CNN) model composed of deep residual network (ResNet) and feature pyramid networks (FPNs) with standard convolution and fully connected heads for mask and box prediction, respectively, was used to automatically depict the choroid layer. The average choroidal thickness and subfoveal choroidal thickness were measured. The results of this study showed that ResNet 50 layers deep (R50) model and ResNet 101 layers deep (R101). R101 U R50 (OR model) demonstrated the best accuracy with an average error of 4.85 pixels and 4.86 pixels, respectively. The R101 ∩ R50 (AND model) took the least time with an average execution time of 4.6 s. Mask-RCNN models showed a good prediction rate of choroidal layer with accuracy rates of 90% and 89.9% for average choroidal thickness and average subfoveal choroidal thickness, respectively. In conclusion, the deep-learning method using the Mask-RCNN model provides a faster and accurate measurement of choroidal thickness. Comparing with manual delineation, it provides better effectiveness, which is feasible for clinical application and larger scale of research on choroid.


Author(s):  
Yun Zhang ◽  
Ling Wang ◽  
Xinqiao Wang ◽  
Chengyun Zhang ◽  
Jiamin Ge ◽  
...  

An effective and rapid deep learning method to predict chemical reactions contributes to the research and development of organic chemistry and drug discovery.


2021 ◽  
Vol 11 (15) ◽  
pp. 7148
Author(s):  
Bedada Endale ◽  
Abera Tullu ◽  
Hayoung Shi ◽  
Beom-Soo Kang

Unmanned aerial vehicles (UAVs) are being widely utilized for various missions: in both civilian and military sectors. Many of these missions demand UAVs to acquire artificial intelligence about the environments they are navigating in. This perception can be realized by training a computing machine to classify objects in the environment. One of the well known machine training approaches is supervised deep learning, which enables a machine to classify objects. However, supervised deep learning comes with huge sacrifice in terms of time and computational resources. Collecting big input data, pre-training processes, such as labeling training data, and the need for a high performance computer for training are some of the challenges that supervised deep learning poses. To address these setbacks, this study proposes mission specific input data augmentation techniques and the design of light-weight deep neural network architecture that is capable of real-time object classification. Semi-direct visual odometry (SVO) data of augmented images are used to train the network for object classification. Ten classes of 10,000 different images in each class were used as input data where 80% were for training the network and the remaining 20% were used for network validation. For the optimization of the designed deep neural network, a sequential gradient descent algorithm was implemented. This algorithm has the advantage of handling redundancy in the data more efficiently than other algorithms.


Author(s):  
Ramesh Adhikari ◽  
Suresh Pokharel

Data augmentation is widely used in image processing and pattern recognition problems in order to increase the richness in diversity of available data. It is commonly used to improve the classification accuracy of images when the available datasets are limited. Deep learning approaches have demonstrated an immense breakthrough in medical diagnostics over the last decade. A significant amount of datasets are needed for the effective training of deep neural networks. The appropriate use of data augmentation techniques prevents the model from over-fitting and thus increases the generalization capability of the network while testing afterward on unseen data. However, it remains a huge challenge to obtain such a large dataset from rare diseases in the medical field. This study presents the synthetic data augmentation technique using Generative Adversarial Networks to evaluate the generalization capability of neural networks using existing data more effectively. In this research, the convolutional neural network (CNN) model is used to classify the X-ray images of the human chest in both normal and pneumonia conditions; then, the synthetic images of the X-ray from the available dataset are generated by using the deep convolutional generative adversarial network (DCGAN) model. Finally, the CNN model is trained again with the original dataset and augmented data generated using the DCGAN model. The classification performance of the CNN model is improved by 3.2% when the augmented data were used along with the originally available dataset.


Sign in / Sign up

Export Citation Format

Share Document