scholarly journals Control the Information of the Image with Anisotropic Diffusion and Isotropic Diffusion for the Image Classification

2021 ◽  
Author(s):  
Hyun-Tae Choi ◽  
Nahyun Lee ◽  
Jewon No ◽  
Sangil Han ◽  
Jaeho Tak ◽  
...  

Humans can recognize objects well even if they only show the shape of objects or an object is composed of several components. But, most of the classifiers in the deep learning framework are trained through original images without removing complex elements inside the object. And also, they do not remove things other than the object to be classified. So the classifiers are not as effective as the human classification of objects because they are trained with the original image which has many objects that the classifier does not want to classify. In this respect, we found out which pre-processing can improve the performance of the classifier the most by comparing the results of using data through other pre-processing. In this paper, we try to limit the amount of information in the object to a minimum. To restrict the information, we use anisotropic diffusion and isotropic diffusion, which are used for removing the noise in the images. By using the anisotropic diffusion and the isotropic diffusion for the pre-processing, only shapes of objects were passed to the classifier. With these diffusion processes, we can get similar classification accuracy compared to when using the original image, and we found out that although the original images are diffused too much, the classifier can classify the objects centered on discriminative parts of the objects.

Author(s):  
Tong Lin ◽  
◽  
Xin Chen ◽  
Xiao Tang ◽  
Ling He ◽  
...  

This paper discusses the use of deep convolutional neural networks for radar target classification. In this paper, three parts of the work are carried out: firstly, effective data enhancement methods are used to augment the dataset and address unbalanced datasets. Second, using deep learning techniques, we explore an effective framework for classifying and identifying targets based on radar spectral map data. By using data enhancement and the framework, we achieved an overall classification accuracy of 0.946. In the end, we researched the automatic annotation of image ROI (region of interest). By adjusting the model, we obtained a 93% accuracy in automatic labeling and classification of targets for both car and cyclist categories.


2019 ◽  
Vol 2019 (1) ◽  
pp. 69-74
Author(s):  
Aldo Barba ◽  
Ivar Farup ◽  
Marius Pedersen

In the paper "Colour-to-Greyscale Image Conversion by Linear Anisotropic Diffusion of Perceptual Colour Metrics", Farup et al. presented an algorithm to convert colour images to greyscale. The algorithm produces greyscale reproductions that preserve detail derived from local colour differences in the original colour image. Such detail is extracted by using linear anisotropic diffusion to build a greyscale reproduction from a gradient of the original image that is in turn calculated using Riemannised colour metrics. The purpose of the current paper is to re-evaluate one of the psychometric experiments for these two methods (CIELAB L* and anisotropic Δ99) by using a flipping method to compare their resulting images instead of the side by side method used in the original evaluation. In addition to testing the two selected algorithms, a third greyscale reproduction was manually created (colour graded) using a colour correction software commonly used to process motion pictures. Results of the psychometric experiment found that when comparing images using the flipping method, there was a statistically significant difference between the anisotropic Δ99 and CIELAB L* conversions that favored the anisotropic method. The comparison between Δ99 conversion and the manually colour graded image also showed a statistically significant difference between them, in this case favoring the colour graded version.


Diagnostics ◽  
2021 ◽  
Vol 11 (2) ◽  
pp. 233
Author(s):  
Dong-Woon Lee ◽  
Sung-Yong Kim ◽  
Seong-Nyum Jeong ◽  
Jae-Hong Lee

Fracture of a dental implant (DI) is a rare mechanical complication that is a critical cause of DI failure and explantation. The purpose of this study was to evaluate the reliability and validity of a three different deep convolutional neural network (DCNN) architectures (VGGNet-19, GoogLeNet Inception-v3, and automated DCNN) for the detection and classification of fractured DI using panoramic and periapical radiographic images. A total of 21,398 DIs were reviewed at two dental hospitals, and 251 intact and 194 fractured DI radiographic images were identified and included as the dataset in this study. All three DCNN architectures achieved a fractured DI detection and classification accuracy of over 0.80 AUC. In particular, automated DCNN architecture using periapical images showed the highest and most reliable detection (AUC = 0.984, 95% CI = 0.900–1.000) and classification (AUC = 0.869, 95% CI = 0.778–0.929) accuracy performance compared to fine-tuned and pre-trained VGGNet-19 and GoogLeNet Inception-v3 architectures. The three DCNN architectures showed acceptable accuracy in the detection and classification of fractured DIs, with the best accuracy performance achieved by the automated DCNN architecture using only periapical images.


Information ◽  
2021 ◽  
Vol 12 (6) ◽  
pp. 249
Author(s):  
Xin Jin ◽  
Yuanwen Zou ◽  
Zhongbing Huang

The cell cycle is an important process in cellular life. In recent years, some image processing methods have been developed to determine the cell cycle stages of individual cells. However, in most of these methods, cells have to be segmented, and their features need to be extracted. During feature extraction, some important information may be lost, resulting in lower classification accuracy. Thus, we used a deep learning method to retain all cell features. In order to solve the problems surrounding insufficient numbers of original images and the imbalanced distribution of original images, we used the Wasserstein generative adversarial network-gradient penalty (WGAN-GP) for data augmentation. At the same time, a residual network (ResNet) was used for image classification. ResNet is one of the most used deep learning classification networks. The classification accuracy of cell cycle images was achieved more effectively with our method, reaching 83.88%. Compared with an accuracy of 79.40% in previous experiments, our accuracy increased by 4.48%. Another dataset was used to verify the effect of our model and, compared with the accuracy from previous results, our accuracy increased by 12.52%. The results showed that our new cell cycle image classification system based on WGAN-GP and ResNet is useful for the classification of imbalanced images. Moreover, our method could potentially solve the low classification accuracy in biomedical images caused by insufficient numbers of original images and the imbalanced distribution of original images.


2021 ◽  
Vol 11 (9) ◽  
pp. 3974
Author(s):  
Laila Bashmal ◽  
Yakoub Bazi ◽  
Mohamad Mahmoud Al Rahhal ◽  
Haikel Alhichri ◽  
Naif Al Ajlan

In this paper, we present an approach for the multi-label classification of remote sensing images based on data-efficient transformers. During the training phase, we generated a second view for each image from the training set using data augmentation. Then, both the image and its augmented version were reshaped into a sequence of flattened patches and then fed to the transformer encoder. The latter extracts a compact feature representation from each image with the help of a self-attention mechanism, which can handle the global dependencies between different regions of the high-resolution aerial image. On the top of the encoder, we mounted two classifiers, a token and a distiller classifier. During training, we minimized a global loss consisting of two terms, each corresponding to one of the two classifiers. In the test phase, we considered the average of the two classifiers as the final class labels. Experiments on two datasets acquired over the cities of Trento and Civezzano with a ground resolution of two-centimeter demonstrated the effectiveness of the proposed model.


Sensors ◽  
2019 ◽  
Vol 19 (4) ◽  
pp. 916 ◽  
Author(s):  
Wen Cao ◽  
Chunmei Liu ◽  
Pengfei Jia

Aroma plays a significant role in the quality of citrus fruits and processed products. The detection and analysis of citrus volatiles can be measured by an electronic nose (E-nose); in this paper, an E-nose is employed to classify the juice which is stored for different days. Feature extraction and classification are two important requirements for an E-nose. During the training process, a classifier can optimize its own parameters to achieve a better classification accuracy but cannot decide its input data which is treated by feature extraction methods, so the classification result is not always ideal. Label consistent KSVD (L-KSVD) is a novel technique which can extract the feature and classify the data at the same time, and such an operation can improve the classification accuracy. We propose an enhanced L-KSVD called E-LCKSVD for E-nose in this paper. During E-LCKSVD, we introduce a kernel function to the traditional L-KSVD and present a new initialization technique of its dictionary; finally, the weighted coefficients of different parts of its object function is studied, and enhanced quantum-behaved particle swarm optimization (EQPSO) is employed to optimize these coefficients. During the experimental section, we firstly find the classification accuracy of KSVD, and L-KSVD is improved with the help of the kernel function; this can prove that their ability of dealing nonlinear data is improved. Then, we compare the results of different dictionary initialization techniques and prove our proposed method is better. Finally, we find the optimal value of the weighted coefficients of the object function of E-LCKSVD that can make E-nose reach a better performance.


Sign in / Sign up

Export Citation Format

Share Document