Optimization-based Framework for Geological Scenario Determination Using Parameterized Training Images

Author(s):  
M.A.H. Rousset ◽  
L.J. Durlofsky
2020 ◽  
Vol 2020 (10) ◽  
pp. 310-1-310-7
Author(s):  
Khalid Omer ◽  
Luca Caucci ◽  
Meredith Kupinski

This work reports on convolutional neural network (CNN) performance on an image texture classification task as a function of linear image processing and number of training images. Detection performance of single and multi-layer CNNs (sCNN/mCNN) are compared to optimal observers. Performance is quantified by the area under the receiver operating characteristic (ROC) curve, also known as the AUC. For perfect detection AUC = 1.0 and AUC = 0.5 for guessing. The Ideal Observer (IO) maximizes AUC but is prohibitive in practice because it depends on high-dimensional image likelihoods. The IO performance is invariant to any fullrank, invertible linear image processing. This work demonstrates the existence of full-rank, invertible linear transforms that can degrade both sCNN and mCNN even in the limit of large quantities of training data. A subsequent invertible linear transform changes the images’ correlation structure again and can improve this AUC. Stationary textures sampled from zero mean and unequal covariance Gaussian distributions allow closed-form analytic expressions for the IO and optimal linear compression. Linear compression is a mitigation technique for high-dimension low sample size (HDLSS) applications. By definition, compression strictly decreases or maintains IO detection performance. For small quantities of training data, linear image compression prior to the sCNN architecture can increase AUC from 0.56 to 0.93. Results indicate an optimal compression ratio for CNN based on task difficulty, compression method, and number of training images.


2021 ◽  
Vol 13 (14) ◽  
pp. 2822
Author(s):  
Zhe Lin ◽  
Wenxuan Guo

An accurate stand count is a prerequisite to determining the emergence rate, assessing seedling vigor, and facilitating site-specific management for optimal crop production. Traditional manual counting methods in stand assessment are labor intensive and time consuming for large-scale breeding programs or production field operations. This study aimed to apply two deep learning models, the MobileNet and CenterNet, to detect and count cotton plants at the seedling stage with unmanned aerial system (UAS) images. These models were trained with two datasets containing 400 and 900 images with variations in plant size and soil background brightness. The performance of these models was assessed with two testing datasets of different dimensions, testing dataset 1 with 300 by 400 pixels and testing dataset 2 with 250 by 1200 pixels. The model validation results showed that the mean average precision (mAP) and average recall (AR) were 79% and 73% for the CenterNet model, and 86% and 72% for the MobileNet model with 900 training images. The accuracy of cotton plant detection and counting was higher with testing dataset 1 for both CenterNet and MobileNet models. The results showed that the CenterNet model had a better overall performance for cotton plant detection and counting with 900 training images. The results also indicated that more training images are required when applying object detection models on images with different dimensions from training datasets. The mean absolute percentage error (MAPE), coefficient of determination (R2), and the root mean squared error (RMSE) values of the cotton plant counting were 0.07%, 0.98 and 0.37, respectively, with testing dataset 1 for the CenterNet model with 900 training images. Both MobileNet and CenterNet models have the potential to accurately and timely detect and count cotton plants based on high-resolution UAS images at the seedling stage. This study provides valuable information for selecting the right deep learning tools and the appropriate number of training images for object detection projects in agricultural applications.


2021 ◽  
pp. 107924
Author(s):  
Wanxuan Lu ◽  
Dong Gong ◽  
Kun Fu ◽  
Xian Sun ◽  
Wenhui Diao ◽  
...  
Keyword(s):  

2021 ◽  
Author(s):  
Khalid Labib Alsamadony ◽  
Ertugrul Umut Yildirim ◽  
Guenther Glatz ◽  
Umair bin Waheed ◽  
Sherif M. Hanafy

Abstract Computed tomography (CT) is an important tool to characterize rock samples allowing quantification of physical properties in 3D and 4D. The accuracy of a property delineated from CT data is strongly correlated with the CT image quality. In general, high-quality, lower noise CT Images mandate greater exposure times. With increasing exposure time, however, more wear is put on the X-Ray tube and longer cooldown periods are required, inevitably limiting the temporal resolution of the particular phenomena under investigation. In this work, we propose a deep convolutional neural network (DCNN) based approach to improve the quality of images collected during reduced exposure time scans. First, we convolve long exposure time images from medical CT scanner with a blur kernel to mimic the degradation caused because of reduced exposure time scanning. Subsequently, utilizing the high- and low-quality scan stacks, we train a DCNN. The trained network enables us to restore any low-quality scan for which high-quality reference is not available. Furthermore, we investigate several factors affecting the DCNN performance such as the number of training images, transfer learning strategies, and loss functions. The results indicate that the number of training images is an important factor since the predictive capability of the DCNN improves as the number of training images increases. We illustrate, however, that the requirement for a large training dataset can be reduced by exploiting transfer learning. In addition, training the DCNN on mean squared error (MSE) as a loss function outperforms both mean absolute error (MAE) and Peak signal-to-noise ratio (PSNR) loss functions with respect to image quality metrics. The presented approach enables the prediction of high-quality images from low exposure CT images. Consequently, this allows for continued scanning without the need for X-Ray tube to cool down, thereby maximizing the temporal resolution. This is of particular value for any core flood experiment seeking to capture the underlying dynamics.


2020 ◽  
Vol 34 (4) ◽  
pp. 387-394
Author(s):  
Soodabeh Amanzadeh ◽  
Yahya Forghani ◽  
Javad Mahdavi Chabok

Kernel extended dictionary learning model (KED) is a new type of Sparse Representation for Classification (SRC), which represents the input face image as a linear combination of dictionary set and extended dictionary set to determine the input face image class label. Extended dictionary is created based on the differences between the occluded images and non-occluded training images. There are four defaults to make about KED: (1) Similar weights are assigned to the principle components of occlusion variations in KED model, while the principle components of the occlusion variations have different weights, which are proportional to the principle components Eigen-values. (2) Reconstruction of an occluded image is not possible by combining only non-occluded images and the principle components (or the directions) of occlusion variations, but it requires the mean of occlusion variations. (3) The importance and capability of main dictionary and extended dictionary in reconstructing the input face image is not the same, necessarily. (4) KED Runtime is high. To address these problems or challenges, a novel mathematical model is proposed in this paper. In the proposed model, different weights are assigned to the principle components of occlusion variations; different weights are assigned to the main dictionary and extended dictionary; an occluded image is reconstructed by non-occluded images and the principle components of occlusion variations, and also the mean of occlusion variations; and collaborative representation is used instead of sparse representation to enhance the runtime. Experimental results on CAS-PEAL subsets showed that the runtime and accuracy of the proposed model is about 1% better than that of KED.


2018 ◽  
Vol 7 (4) ◽  
pp. 9 ◽  
Author(s):  
Shakir F. Kak ◽  
Firas M. Mustafa ◽  
Pedro R. Valente

In a recent past, face recognition was one of the most popular methods and successful application of image processing field which is widely used in security and biometric applications. The innovation of new approaches to face identification technologies is continuously subject to building much strong face recognition algorithms. Face recognition in real-time applications has been fast-growing challenging and interesting. The human face identification process is not trivial task especially different face lighting and poses are captured to be matched. In this study, the proposed method is tested using a benchmark ORL database that contains 400 images of 40 persons as the variant posse, lighting, etc. Discrete avelet Transform technique is applied on the ORL database to enhance the accuracy and the recognition rate. The best recognition rate result obtained is 99.25%, when tested using 9 training images and 1 testing image with cosine distance measurement. The recognition rate Increased when applying 2-level of DWT with the bior5.5 filter on training image database and the test image. For feature extraction and dimension reduction, PCA is used. Euclidean distance, Manhattan distance, and Cosine distance are Distance measures used for the matching process.


1996 ◽  
Vol 35 (2) ◽  
pp. 314
Author(s):  
J. D. Brasher ◽  
Mark Woodson

Sign in / Sign up

Export Citation Format

Share Document