Element extraction and convolutional neural network-based classification for blue calico

2020 ◽  
pp. 004051752093957
Author(s):  
Xiaojun Jia ◽  
Zihao Liu

Blue calico is a highly valued folk handicraft that forms part of China’s national intangible cultural heritage. Thus, blue calico is a worthy target for reconstruction using modern image processing technology. Extracting the visual components or elements of a blue calico pattern is one way to capture the underlying design and enable innovation in traditional patterns using modern techniques. This paper presents a method of element extraction and classification based on a smart convolutional neural network (CNN), with an improved CifarNet structure, which we call CalicoNet. Initially, the algorithm for element extraction is implemented to generate element samples of blue calico. This process includes gray scaling, binarization, and contour extraction. We construct a data set of elements with 12 types. Then, four critical hyper-parameters, the batch-size, dropout ratio, learning rate, and pooling strategy, are optimized by a comparative analysis. A combination classifier strategy is subsequently added to the fully connected layers to strengthen the feature expression in the corresponding classes. Finally, the superiority of the proposed CalicoNet is verified through a comparison with other sophisticated CNNs. Experimental results demonstrate that CalicoNet achieves a validation accuracy of 99.2% for the training set, a total time of 1.13 hours for the whole data set, and a test mean accuracy precision of 98.66%. The robust performance of the proposed method across the element data set indicates that CalicoNet is a promising approach for element extraction and classification.

Sensors ◽  
2019 ◽  
Vol 19 (19) ◽  
pp. 4161 ◽  
Author(s):  
Hang ◽  
Zhang ◽  
Chen ◽  
Zhang ◽  
Wang

Plant leaf diseases are closely related to people's daily life. Due to the wide variety of diseases, it is not only time-consuming and labor-intensive to identify and classify diseases by artificial eyes, but also easy to be misidentified with having a high error rate. Therefore, we proposed a deep learning-based method to identify and classify plant leaf diseases. The proposed method can take the advantages of the neural network to extract the characteristics of diseased parts, and thus to classify target disease areas. To address the issues of long training convergence time and too-large model parameters, the traditional convolutional neural network was improved by combining a structure of inception module, a squeeze-and-excitation (SE) module and a global pooling layer to identify diseases. Through the Inception structure, the feature data of the convolutional layer were fused in multi-scales to improve the accuracy on the leaf disease dataset. Finally, the global average pooling layer was used instead of the fully connected layer to reduce the number of model parameters. Compared with some traditional convolutional neural networks, our model yielded better performance and achieved an accuracy of 91.7% on the test data set. At the same time, the number of model parameters and training time have also been greatly reduced. The experimental classification on plant leaf diseases indicated that our method is feasible and effective.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Haoyan Yang ◽  
Jiangong Ni ◽  
Jiyue Gao ◽  
Zhongzhi Han ◽  
Tao Luan

AbstractCrop variety identification is an essential link in seed detection, phenotype collection and scientific breeding. This paper takes peanut as an example to explore a new method for crop variety identification. Peanut is a crucial oil crop and cash crop. The yield and quality of different peanut varieties are different, so it is necessary to identify and classify different peanut varieties. The traditional image processing method of peanut variety identification needs to extract many features, which has defects such as intense subjectivity and insufficient generalization ability. Based on the deep learning technology, this paper improved the deep convolutional neural network VGG16 and applied the improved VGG16 to the identification and classification task of 12 varieties of peanuts. Firstly, the peanut pod images of 12 varieties obtained by the scanner were preprocessed with gray-scale, binarization, and ROI extraction to form a peanut pod data set with a total of 3365 images of 12 varieties. A series of improvements have been made to VGG16. Remove the F6 and F7 fully connected layers of VGG16. Add Conv6 and Global Average Pooling Layer. The three convolutional layers of conv5 have changed into Depth Concatenation and add the Batch Normalization(BN) layers to the model. Besides, fine-tuning is carried out based on the improved VGG16. We adjusted the location of the BN layers. Adjust the number of filters for Conv6. Finally, the improved VGG16 model's training test results were compared with the other classic models, AlexNet, VGG16, GoogLeNet, ResNet18, ResNet50, SqueezeNet, DenseNet201 and MobileNetv2 verify its superiority. The average accuracy of the improved VGG16 model on the peanut pods test set was 96.7%, which was 8.9% higher than that of VGG16, and 1.6–12.3% higher than that of other classical models. Besides, supplementary experiments were carried out to prove the robustness and generality of the improved VGG16. The improved VGG16 was applied to the identification and classification of seven corn grain varieties with the same method and an average accuracy of 90.1% was achieved. The experimental results show that the improved VGG16 proposed in this paper can identify and classify peanut pods of different varieties, proving the feasibility of a convolutional neural network in variety identification and classification. The model proposed in this experiment has a positive significance for exploring other Crop variety identification and classification.


Information ◽  
2019 ◽  
Vol 10 (5) ◽  
pp. 162 ◽  
Author(s):  
Jiana Meng ◽  
Yingchun Long ◽  
Yuhai Yu ◽  
Dandan Zhao ◽  
Shuang Liu

Transfer learning is one of the popular methods for solving the problem that the models built on the source domain cannot be directly applied to the target domain in the cross-domain sentiment classification. This paper proposes a transfer learning method based on the multi-layer convolutional neural network (CNN). Interestingly, we construct a convolutional neural network model to extract features from the source domain and share the weights in the convolutional layer and the pooling layer between the source and target domain samples. Next, we fine-tune the weights in the last layer, named the fully connected layer, and transfer the models from the source domain to the target domain. Comparing with the classical transfer learning methods, the method proposed in this paper does not need to retrain the network for the target domain. The experimental evaluation of the cross-domain data set shows that the proposed method achieves a relatively good performance.


Circulation ◽  
2018 ◽  
Vol 138 (Suppl_2) ◽  
Author(s):  
Tetsuo Hatanaka ◽  
Hiroshi Kaneko ◽  
Aki Nagase ◽  
Seishiro Marukawa

Introduction: An interruption of chest compressions during CPR adversely affects patient outcome. Currently, however, periodical interruptions are unavoidable to assess the ECG rhythms and to give shocks for defibrillation if indicated. Evidence suggests a 5-second interruption immediately before shocks may translate into ~15% reduction of the chance of survival. The objective of this study was to build, train and validate a convolutional neural network (artificial intelligence) for detecting shock-indicated rhythms out of ECG signals corrupted with chest compression artifacts during CPR. Methods: Our convolutional neural network consisted of 7 convolutional layers, 3 pooling layers and 3 fully-connected layers for binary classification (shock-indicated vs non-shock-indicated). The input data set was a spectrogram consisting of 56 frequency-bins by 80 time-segments transformed from a 12.16-seconds ECG signal. From AEDs used for 236 patients with out-of-hospital cardiac arrest, 1,223 annotated ECG strips were extracted. Ventricular fibrillation and wide-QRS ventricular tachycardia with HR>180 beats/min were annotated as shock-indicated, and the others as non-shock-indicated. The total length of the strips was 8:49:57 (hr:min:sec) and 8:02:07 respectively for shock-indicated and non-shock-indicated rhythms. Those strips were converted into 465,102 spectrograms allowing partial overlaps and were fed into the neural network for training. The validation data set was obtained from a separate group of 225 patients, from which annotated ECG strips (total duration of 62:11:28) were extracted, yielding 43,800 spectrograms. Results: After the training, both the sensitivity and specificity of detecting shock-indicated rhythms over the training data set were 99.7% - 100% (varying with training instances). The sensitivity and specificity over the validation data set were 99.3% - 99.7% and 99.3% - 99.5%, respectively. Conclusions: The convolutional neural network has accurately and continuously evaluated the ECG rhythms during CPR, potentially obviating the need for rhythm checks for defibrillation during CPR.


2019 ◽  
Vol 24 (3) ◽  
pp. 220-228
Author(s):  
Gusti Alfahmi Anwar ◽  
Desti Riminarsih

Panthera merupakan genus dari keluarga kucing yang memiliki empat spesies popular yaitu, harimau, jaguar, macan tutul, singa. Singa memiliki warna keemasan dan tidak memilki motif, harimau memiliki motif loreng dengan garis-garis panjang, jaguar memiliki tubuh yang lebih besar dari pada macan tutul serta memiliki motif tutul yang lebih lebar, sedangkan macan tutul memiliki tubuh yang sedikit lebih ramping dari pada jaguar dan memiliki tutul yang tidak terlalu lebar. Pada penelitian ini dilakukan klasifikasi genus panther yaitu harimau, jaguar, macan tutul, dan singa menggunakan metode Convolutional Neural Network. Model Convolutional Neural Network yang digunakan memiliki 1 input layer, 5 convolution layer, dan 2 fully connected layer. Dataset yang digunakan berupa citra harimau, jaguar, macan tutul, dan singa. Data training terdiri dari 3840 citra, data validasi sebanyak 960 citra, dan data testing sebanyak 800 citra. Hasil akurasi dari pelatihan model untuk training yaitu 92,31% dan validasi yaitu 81,88%, pengujian model menggunakan dataset testing mendapatan hasil 68%. Hasil akurasi prediksi didapatkan dari nilai F1-Score pada pengujian didapatkan sebesar 78% untuk harimau, 70% untuk jaguar, 37% untuk macan tutul, 74% untuk singa. Macan tutul mendapatkan akurasi terendah dibandingkan 3 hewan lainnya tetapi lebih baik dibandingkan hasil penelitian sebelumnya.


2021 ◽  
Vol 7 (2) ◽  
pp. 356-362
Author(s):  
Harry Coppock ◽  
Alex Gaskell ◽  
Panagiotis Tzirakis ◽  
Alice Baird ◽  
Lyn Jones ◽  
...  

BackgroundSince the emergence of COVID-19 in December 2019, multidisciplinary research teams have wrestled with how best to control the pandemic in light of its considerable physical, psychological and economic damage. Mass testing has been advocated as a potential remedy; however, mass testing using physical tests is a costly and hard-to-scale solution.MethodsThis study demonstrates the feasibility of an alternative form of COVID-19 detection, harnessing digital technology through the use of audio biomarkers and deep learning. Specifically, we show that a deep neural network based model can be trained to detect symptomatic and asymptomatic COVID-19 cases using breath and cough audio recordings.ResultsOur model, a custom convolutional neural network, demonstrates strong empirical performance on a data set consisting of 355 crowdsourced participants, achieving an area under the curve of the receiver operating characteristics of 0.846 on the task of COVID-19 classification.ConclusionThis study offers a proof of concept for diagnosing COVID-19 using cough and breath audio signals and motivates a comprehensive follow-up research study on a wider data sample, given the evident advantages of a low-cost, highly scalable digital COVID-19 diagnostic tool.


Entropy ◽  
2021 ◽  
Vol 23 (1) ◽  
pp. 119
Author(s):  
Tao Wang ◽  
Changhua Lu ◽  
Yining Sun ◽  
Mei Yang ◽  
Chun Liu ◽  
...  

Early detection of arrhythmia and effective treatment can prevent deaths caused by cardiovascular disease (CVD). In clinical practice, the diagnosis is made by checking the electrocardiogram (ECG) beat-by-beat, but this is usually time-consuming and laborious. In the paper, we propose an automatic ECG classification method based on Continuous Wavelet Transform (CWT) and Convolutional Neural Network (CNN). CWT is used to decompose ECG signals to obtain different time-frequency components, and CNN is used to extract features from the 2D-scalogram composed of the above time-frequency components. Considering the surrounding R peak interval (also called RR interval) is also useful for the diagnosis of arrhythmia, four RR interval features are extracted and combined with the CNN features to input into a fully connected layer for ECG classification. By testing in the MIT-BIH arrhythmia database, our method achieves an overall performance of 70.75%, 67.47%, 68.76%, and 98.74% for positive predictive value, sensitivity, F1-score, and accuracy, respectively. Compared with existing methods, the overall F1-score of our method is increased by 4.75~16.85%. Because our method is simple and highly accurate, it can potentially be used as a clinical auxiliary diagnostic tool.


2020 ◽  
Vol 2020 ◽  
pp. 1-6
Author(s):  
Jian-ye Yuan ◽  
Xin-yuan Nan ◽  
Cheng-rong Li ◽  
Le-le Sun

Considering that the garbage classification is urgent, a 23-layer convolutional neural network (CNN) model is designed in this paper, with the emphasis on the real-time garbage classification, to solve the low accuracy of garbage classification and recycling and difficulty in manual recycling. Firstly, the depthwise separable convolution was used to reduce the Params of the model. Then, the attention mechanism was used to improve the accuracy of the garbage classification model. Finally, the model fine-tuning method was used to further improve the performance of the garbage classification model. Besides, we compared the model with classic image classification models including AlexNet, VGG16, and ResNet18 and lightweight classification models including MobileNetV2 and SuffleNetV2 and found that the model GAF_dense has a higher accuracy rate, fewer Params, and FLOPs. To further check the performance of the model, we tested the CIFAR-10 data set and found the accuracy rates of the model (GAF_dense) are 0.018 and 0.03 higher than ResNet18 and SufflenetV2, respectively. In the ImageNet data set, the accuracy rates of the model (GAF_dense) are 0.225 and 0.146 higher than Resnet18 and SufflenetV2, respectively. Therefore, the garbage classification model proposed in this paper is suitable for garbage classification and other classification tasks to protect the ecological environment, which can be applied to classification tasks such as environmental science, children’s education, and environmental protection.


2020 ◽  
pp. 1-11
Author(s):  
Jie Liu ◽  
Hongbo Zhao

BACKGROUND: Convolution neural network is often superior to other similar algorithms in image classification. Convolution layer and sub-sampling layer have the function of extracting sample features, and the feature of sharing weights greatly reduces the training parameters of the network. OBJECTIVE: This paper describes the improved convolution neural network structure, including convolution layer, sub-sampling layer and full connection layer. This paper also introduces five kinds of diseases and normal eye images reflected by the blood filament of the eyeball “yan.mat” data set, convenient to use MATLAB software for calculation. METHODSL: In this paper, we improve the structure of the classical LeNet-5 convolutional neural network, and design a network structure with different convolution kernels, different sub-sampling methods and different classifiers, and use this structure to solve the problem of ocular bloodstream disease recognition. RESULTS: The experimental results show that the improved convolutional neural network structure is ideal for the recognition of eye blood silk data set, which shows that the convolution neural network has the characteristics of strong classification and strong robustness. The improved structure can classify the diseases reflected by eyeball bloodstain well.


Sign in / Sign up

Export Citation Format

Share Document