scholarly journals Airborne Network Traffic Identification Method under Small Training Samples

Author(s):  
Na Lyu ◽  
Jiaxin Zhou ◽  
Zhuo Chen ◽  
Wu Chen

Due to the high cost and difficulty of traffic data set acquisition and the high time sensitivity of traffic distribution, the machine learning-based traffic identification method is difficult to be applied in airborne network environment. Aiming at this problem, a method for airborne network traffic identification based on the convolutional neural network under small traffic samples is proposed. Firstly, the pre-training of the initial model for the convolutional neural network is implemented based on the complete data set in source domain, and then the retraining of the convolutional neural network is realized through the layer frozen based fine-tuning learning algorithm of convolutional neural network on the incomplete dataset in target domain, and the convolutional neural network model based feature representing transferring(FRT-CNN) is constructed to realize online traffic identification. The experiment results on the actual airborne network traffic dataset show that the proposed method can guarantee the accuracy of traffic identification under limited traffic samples, and the classification performance is significantly improved comparing with the existing small-sample learning methods.

2020 ◽  
Vol 2020 ◽  
pp. 1-6
Author(s):  
Jian-ye Yuan ◽  
Xin-yuan Nan ◽  
Cheng-rong Li ◽  
Le-le Sun

Considering that the garbage classification is urgent, a 23-layer convolutional neural network (CNN) model is designed in this paper, with the emphasis on the real-time garbage classification, to solve the low accuracy of garbage classification and recycling and difficulty in manual recycling. Firstly, the depthwise separable convolution was used to reduce the Params of the model. Then, the attention mechanism was used to improve the accuracy of the garbage classification model. Finally, the model fine-tuning method was used to further improve the performance of the garbage classification model. Besides, we compared the model with classic image classification models including AlexNet, VGG16, and ResNet18 and lightweight classification models including MobileNetV2 and SuffleNetV2 and found that the model GAF_dense has a higher accuracy rate, fewer Params, and FLOPs. To further check the performance of the model, we tested the CIFAR-10 data set and found the accuracy rates of the model (GAF_dense) are 0.018 and 0.03 higher than ResNet18 and SufflenetV2, respectively. In the ImageNet data set, the accuracy rates of the model (GAF_dense) are 0.225 and 0.146 higher than Resnet18 and SufflenetV2, respectively. Therefore, the garbage classification model proposed in this paper is suitable for garbage classification and other classification tasks to protect the ecological environment, which can be applied to classification tasks such as environmental science, children’s education, and environmental protection.


2020 ◽  
Vol 102-B (6_Supple_A) ◽  
pp. 101-106
Author(s):  
Romil F. Shah ◽  
Stefano A. Bini ◽  
Alejandro M. Martinez ◽  
Valentina Pedoia ◽  
Thomas P. Vail

Aims The aim of this study was to evaluate the ability of a machine-learning algorithm to diagnose prosthetic loosening from preoperative radiographs and to investigate the inputs that might improve its performance. Methods A group of 697 patients underwent a first-time revision of a total hip (THA) or total knee arthroplasty (TKA) at our institution between 2012 and 2018. Preoperative anteroposterior (AP) and lateral radiographs, and historical and comorbidity information were collected from their electronic records. Each patient was defined as having loose or fixed components based on the operation notes. We trained a series of convolutional neural network (CNN) models to predict a diagnosis of loosening at the time of surgery from the preoperative radiographs. We then added historical data about the patients to the best performing model to create a final model and tested it on an independent dataset. Results The convolutional neural network we built performed well when detecting loosening from radiographs alone. The first model built de novo with only the radiological image as input had an accuracy of 70%. The final model, which was built by fine-tuning a publicly available model named DenseNet, combining the AP and lateral radiographs, and incorporating information from the patient’s history, had an accuracy, sensitivity, and specificity of 88.3%, 70.2%, and 95.6% on the independent test dataset. It performed better for cases of revision THA with an accuracy of 90.1%, than for cases of revision TKA with an accuracy of 85.8%. Conclusion This study showed that machine learning can detect prosthetic loosening from radiographs. Its accuracy is enhanced when using highly trained public algorithms, and when adding clinical data to the algorithm. While this algorithm may not be sufficient in its present state of development as a standalone metric of loosening, it is currently a useful augment for clinical decision making. Cite this article: Bone Joint J 2020;102-B(6 Supple A):101–106.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Haoyan Yang ◽  
Jiangong Ni ◽  
Jiyue Gao ◽  
Zhongzhi Han ◽  
Tao Luan

AbstractCrop variety identification is an essential link in seed detection, phenotype collection and scientific breeding. This paper takes peanut as an example to explore a new method for crop variety identification. Peanut is a crucial oil crop and cash crop. The yield and quality of different peanut varieties are different, so it is necessary to identify and classify different peanut varieties. The traditional image processing method of peanut variety identification needs to extract many features, which has defects such as intense subjectivity and insufficient generalization ability. Based on the deep learning technology, this paper improved the deep convolutional neural network VGG16 and applied the improved VGG16 to the identification and classification task of 12 varieties of peanuts. Firstly, the peanut pod images of 12 varieties obtained by the scanner were preprocessed with gray-scale, binarization, and ROI extraction to form a peanut pod data set with a total of 3365 images of 12 varieties. A series of improvements have been made to VGG16. Remove the F6 and F7 fully connected layers of VGG16. Add Conv6 and Global Average Pooling Layer. The three convolutional layers of conv5 have changed into Depth Concatenation and add the Batch Normalization(BN) layers to the model. Besides, fine-tuning is carried out based on the improved VGG16. We adjusted the location of the BN layers. Adjust the number of filters for Conv6. Finally, the improved VGG16 model's training test results were compared with the other classic models, AlexNet, VGG16, GoogLeNet, ResNet18, ResNet50, SqueezeNet, DenseNet201 and MobileNetv2 verify its superiority. The average accuracy of the improved VGG16 model on the peanut pods test set was 96.7%, which was 8.9% higher than that of VGG16, and 1.6–12.3% higher than that of other classical models. Besides, supplementary experiments were carried out to prove the robustness and generality of the improved VGG16. The improved VGG16 was applied to the identification and classification of seven corn grain varieties with the same method and an average accuracy of 90.1% was achieved. The experimental results show that the improved VGG16 proposed in this paper can identify and classify peanut pods of different varieties, proving the feasibility of a convolutional neural network in variety identification and classification. The model proposed in this experiment has a positive significance for exploring other Crop variety identification and classification.


2021 ◽  
Vol 11 ◽  
Author(s):  
Yinxiang Guo ◽  
Jianing Xu ◽  
Xiangzhi Li ◽  
Lin Zheng ◽  
Wei Pan ◽  
...  

Patients with thyroid cancer will take a small dose of 131I after undergoing a total thyroidectomy. Single-photon emission computed tomography (SPECT) is used to diagnose whether thyroid tissue remains in the body. However, it is difficult for human eyes to observe the specificity of SPECT images in different categories, and it is difficult for doctors to accurately diagnose the residual thyroid tissue in patients based on SPECT images. At present, the research on the classification of thyroid tissue residues after thyroidectomy is still in a blank state. This paper proposes a ResNet-18 fine-tuning method based on the convolutional neural network model. First, preprocess the SPECT images to improve the image quality and remove background interference. Secondly, use the preprocessed image samples to fine-tune the pretrained ResNet-18 model to obtain better features and finally use the Softmax classifier to diagnose the residual thyroid tissue. The method has been tested on SPECT images of 446 patients collected by local hospital and compared with the widely used lightweight network SqueezeNet model and ShuffleNetV2 model. Due to the small data set, this paper conducted 10 random grouping experiments. Each experiment divided the data set into training set and test set at a ratio of 3:1. The accuracy and sensitivity rates of the model proposed in this paper are 96.69% and 94.75%, which are significantly higher than other models (p < 0.05). The specificity and precision rates are 99.6% and 99.96%, respectively, and there is no significant difference compared with other models. (p > 0.05). The area under the curve of the proposed model, SqueezeNet, and ShuffleNetv2 are 0.988 (95% CI, 0.941–1.000), 0.898 (95% CI, 0.819–0.951) (p = 0.0257), and 0.885 (95% CI, 0.803–0.941) (p = 0.0057) (p < 0.05). We prove that this thyroid tissue residue classification system can be used as a computer-aided diagnosis method to effectively improve the diagnostic accuracy of thyroid tissue residues. While more accurately diagnosing patients with residual thyroid tissue in the body, we try our best to avoid the occurrence of overtreatment, which reflects its potential clinical application value.


Author(s):  
В’ячеслав Васильович Москаленко ◽  
Альона Сергіївна Москаленко ◽  
Артем Геннадійович Коробов ◽  
Микола Олександрович Зарецький ◽  
Віктор Анатолійович Семашко

The efficient model and learning algorithm of the small object detection system for compact aerial vehicle under conditions of restricted computing resources and the limited volume of the labeled learning set are developed. The four-stage learning algorithm of the object detector is proposed. At the first stage, selecting the type of deep convolutional neural network and the number of low-level layers that is pretrained on the ImageNet dataset for reusing takes place. The second stage involves unsupervised learning of high-level convolutional sparse coding layers using the modification of growing neural gas to automatically determine the required number of neurons and provide optimal distributions of the neurons over the data. Its application makes it possible to utilize the unlabeled learning datasets for the adaptation of the high-level feature description to the domain application area. At the third stage, the output feature map is formed by concatenation of feature maps from the different level of the deep convolutional neural network. At that, there is a reduction of output feature map using principal component analysis and followed by the building of decision rules. In order to perform the classification analysis of output, feature map is proposed to use information-extreme classifier learning on principles of boosting. Besides that, the orthogonal incremental extreme learning machine is used to build the regression model for the predict bounding box of the detected small object. The last stage involves fine-tuning of high-level layers of deep network using simulated annealing metaheuristic algorithm in order to approximate the global optimum of the complex criterion of learning efficiency of detection model. As a result of the use of proposed approach has been achieved 96% correctly detection of objects on the images of the open test dataset which indicates the suitability of the model and learning algorithm for practical use. In this case, the size of the learning dataset that has been used to construct the model was 500 unlabeled and 200 labeled learning samples


2021 ◽  
Vol 2021 ◽  
pp. 1-11
Author(s):  
Bin Zheng ◽  
Tao Huang

In order to achieve the accuracy of mango grading, a mango grading system was designed by using the deep learning method. The system mainly includes CCD camera image acquisition, image preprocessing, model training, and model evaluation. Aiming at the traditional deep learning, neural network training needs a large number of sample data sets; a convolutional neural network is proposed to realize the efficient grading of mangoes through the continuous adjustment and optimization of super-parameters and batch size. The ultra-lightweight SqueezeNet related algorithm is introduced. Compared with AlexNet and other related algorithms with the same accuracy level, it has the advantages of small model scale and fast operation speed. The experimental results show that the convolutional neural network model after super-parameters optimization and adjustment has excellent effect on deep learning image processing of small sample data set. Two hundred thirty-four Jinhuang mangoes of Panzhihua were picked in the natural environment and tested. The analysis results can meet the requirements of the agricultural industry standard of the People’s Republic of China—mango and mango grade specification. At the same time, the average accuracy rate was 97.37%, the average error rate was 2.63%, and the average loss value of the model was 0.44. The processing time of an original image with a resolution of 500 × 374 was only 2.57 milliseconds. This method has important theoretical and application value and can provide a powerful means for mango automatic grading.


Information ◽  
2019 ◽  
Vol 10 (5) ◽  
pp. 162 ◽  
Author(s):  
Jiana Meng ◽  
Yingchun Long ◽  
Yuhai Yu ◽  
Dandan Zhao ◽  
Shuang Liu

Transfer learning is one of the popular methods for solving the problem that the models built on the source domain cannot be directly applied to the target domain in the cross-domain sentiment classification. This paper proposes a transfer learning method based on the multi-layer convolutional neural network (CNN). Interestingly, we construct a convolutional neural network model to extract features from the source domain and share the weights in the convolutional layer and the pooling layer between the source and target domain samples. Next, we fine-tune the weights in the last layer, named the fully connected layer, and transfer the models from the source domain to the target domain. Comparing with the classical transfer learning methods, the method proposed in this paper does not need to retrain the network for the target domain. The experimental evaluation of the cross-domain data set shows that the proposed method achieves a relatively good performance.


Author(s):  
Dr. Abhay E Wagh

Abstract: Now a day, with the rapid advancement in the digital contents identification, auto classification of the images is most challenging job in the computer field. Programmed comprehension and breaking down of pictures by framework is troublesome when contrasted with human visions. A Several research have been done to defeat issue in existing classification system,, yet the yield was limited distinctly to low even out picture natives. Nonetheless, those approach need with exact order of pictures. This system uses deep learning algorithm concept to achieve the desired results in this area like computer. Our framework presents Convolutional Neural Network (CNN), a machine learning algorithm is used for automatic classification the images. This system uses the Digit of MNIST data set as a bench mark for classification of gray-scale images. The gray-scale images are used for training which requires more computational power for classification of those images. Using CNN network the result is near about 98% accuracy. Our model accomplishes the high precision in grouping of images. Keywords: Convolutional Neural Network (CNN), deep learning, MINIST, Machine Learning.


Sign in / Sign up

Export Citation Format

Share Document