Gaussian pyramid based laws' mask descriptor for texture classification

Author(s):  
Sonali Dash ◽  
Uma Ranjan Jena
2020 ◽  
Vol 2020 (10) ◽  
pp. 310-1-310-7
Author(s):  
Khalid Omer ◽  
Luca Caucci ◽  
Meredith Kupinski

This work reports on convolutional neural network (CNN) performance on an image texture classification task as a function of linear image processing and number of training images. Detection performance of single and multi-layer CNNs (sCNN/mCNN) are compared to optimal observers. Performance is quantified by the area under the receiver operating characteristic (ROC) curve, also known as the AUC. For perfect detection AUC = 1.0 and AUC = 0.5 for guessing. The Ideal Observer (IO) maximizes AUC but is prohibitive in practice because it depends on high-dimensional image likelihoods. The IO performance is invariant to any fullrank, invertible linear image processing. This work demonstrates the existence of full-rank, invertible linear transforms that can degrade both sCNN and mCNN even in the limit of large quantities of training data. A subsequent invertible linear transform changes the images’ correlation structure again and can improve this AUC. Stationary textures sampled from zero mean and unequal covariance Gaussian distributions allow closed-form analytic expressions for the IO and optimal linear compression. Linear compression is a mitigation technique for high-dimension low sample size (HDLSS) applications. By definition, compression strictly decreases or maintains IO detection performance. For small quantities of training data, linear image compression prior to the sCNN architecture can increase AUC from 0.56 to 0.93. Results indicate an optimal compression ratio for CNN based on task difficulty, compression method, and number of training images.


2020 ◽  
Vol 64 (2) ◽  
pp. 20506-1-20506-7
Author(s):  
Min Zhu ◽  
Rongfu Zhang ◽  
Pei Ma ◽  
Xuedian Zhang ◽  
Qi Guo

Abstract Three-dimensional (3D) reconstruction is extensively used in microscopic applications. Reducing excessive error points and achieving accurate matching of weak texture regions have been the classical challenges for 3D microscopic vision. A Multi-ST algorithm was proposed to improve matching accuracy. The process is performed in two main stages: scaled microscopic images and regularized cost aggregation. First, microscopic image pairs with different scales were extracted according to the Gaussian pyramid criterion. Second, a novel cost aggregation approach based on the regularized multi-scale model was implemented into all scales to obtain the final cost. To evaluate the performances of the proposed Multi-ST algorithm and compare different algorithms, seven groups of images from the Middlebury dataset and four groups of experimental images obtained by a binocular microscopic system were analyzed. Disparity maps and reconstruction maps generated by the proposed approach contained more information and fewer outliers or artifacts. Furthermore, 3D reconstruction of the plug gauges using the Multi-ST algorithm showed that the error was less than 0.025 mm.


2014 ◽  
Vol 1 (3) ◽  
pp. 23-31
Author(s):  
Basava Raju ◽  
◽  
K. Y. Rama Devi ◽  
P. V. Kumar ◽  
◽  
...  

2016 ◽  
Author(s):  
Zilong Zou ◽  
Jie Yang ◽  
Vasileios Megalooikonomou ◽  
Rachid Jennane ◽  
Erkang Cheng ◽  
...  

2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Lei Yan ◽  
Qun Hao ◽  
Jie Cao ◽  
Rizvi Saad ◽  
Kun Li ◽  
...  

AbstractImage fusion integrates information from multiple images (of the same scene) to generate a (more informative) composite image suitable for human and computer vision perception. The method based on multiscale decomposition is one of the commonly fusion methods. In this study, a new fusion framework based on the octave Gaussian pyramid principle is proposed. In comparison with conventional multiscale decomposition, the proposed octave Gaussian pyramid framework retrieves more information by decomposing an image into two scale spaces (octave and interval spaces). Different from traditional multiscale decomposition with one set of detail and base layers, the proposed method decomposes an image into multiple sets of detail and base layers, and it efficiently retains high- and low-frequency information from the original image. The qualitative and quantitative comparison with five existing methods (on publicly available image databases) demonstrate that the proposed method has better visual effects and scores the highest in objective evaluation.


Sensors ◽  
2021 ◽  
Vol 21 (3) ◽  
pp. 1010
Author(s):  
Claudio Cusano ◽  
Paolo Napoletano ◽  
Raimondo Schettini

In this paper we present T1K+, a very large, heterogeneous database of high-quality texture images acquired under variable conditions. T1K+ contains 1129 classes of textures ranging from natural subjects to food, textile samples, construction materials, etc. T1K+ allows the design of experiments especially aimed at understanding the specific issues related to texture classification and retrieval. To help the exploration of the database, all the 1129 classes are hierarchically organized in 5 thematic categories and 266 sub-categories. To complete our study, we present an evaluation of hand-crafted and learned visual descriptors in supervised texture classification tasks.


Sign in / Sign up

Export Citation Format

Share Document