scholarly journals Improving enzyme optimum temperature prediction with resampling strategies and ensemble learning

2020 ◽  
Author(s):  
Japheth E. Gado ◽  
Gregg T. Beckham ◽  
Christina M. Payne

ABSTRACTAccurate prediction of the optimal catalytic temperature (Topt) of enzymes is vital in biotechnology, as enzymes with high Topt values are desired for enhanced reaction rates. Recently, a machine-learning method (TOME) for predicting Topt was developed. TOME was trained on a normally-distributed dataset with a median Topt of 37°C and less than five percent of Topt values above 85°C, limiting the method’s predictive capabilities for thermostable enzymes. Due to the distribution of the training data, the mean squared error on Topt values greater than 85°C is nearly an order of magnitude higher than the error on values between 30 and 50°C. In this study, we apply ensemble learning and resampling strategies that tackle the data imbalance to significantly decrease the error on high Topt values (>85°C) by 60% and increase the overall R2 value from 0.527 to 0.632. The revised method, TOMER, and the resampling strategies applied in this work are freely available to other researchers as a Python package on GitHub.

2021 ◽  
Vol 1 (2) ◽  
pp. 54-58
Author(s):  
Ninta Liana Br Sitepu

Backpropagationcial neural networks are one of the artificial representations of the human brain that are always trying to stimulate the learning process of the human brain. Backpropagation is a gradient descent method to minimize the squared of the output error. Backprorpagation works through an iterative process using a set of sample data (training data), comparing the predicted value of the network with each sample data. In each process, the weight of the relation in the network is modified to minimize the Mean Squared Error value between the predicted value from the network and the actual value. The purpose of this thesis is to be able to help teachers at SMP Negeri 1 Salakaran to predict the value of student learning. In the calculation using the maximum epouch = 10000, the target error is 0.01, and the learning rate is 0.3, then there is a calculation result where the need ratio A has a value of 0.7517, which means that the value has decreased and D has a value of 0.9202 which means that this value has increased..


2011 ◽  
Vol 60 (2) ◽  
pp. 248-255 ◽  
Author(s):  
Sangmun Shin ◽  
Funda Samanlioglu ◽  
Byung Rae Cho ◽  
Margaret M. Wiecek

2018 ◽  
Vol 10 (12) ◽  
pp. 4863 ◽  
Author(s):  
Chao Huang ◽  
Longpeng Cao ◽  
Nanxin Peng ◽  
Sijia Li ◽  
Jing Zhang ◽  
...  

Photovoltaic (PV) modules convert renewable and sustainable solar energy into electricity. However, the uncertainty of PV power production brings challenges for the grid operation. To facilitate the management and scheduling of PV power plants, forecasting is an essential technique. In this paper, a robust multilayer perception (MLP) neural network was developed for day-ahead forecasting of hourly PV power. A generic MLP is usually trained by minimizing the mean squared loss. The mean squared error is sensitive to a few particularly large errors that can lead to a poor estimator. To tackle the problem, the pseudo-Huber loss function, which combines the best properties of squared loss and absolute loss, was adopted in this paper. The effectiveness and efficiency of the proposed method was verified by benchmarking against a generic MLP network with real PV data. Numerical experiments illustrated that the proposed method performed better than the generic MLP network in terms of root mean squared error (RMSE) and mean absolute error (MAE).


2016 ◽  
Vol 5 (1) ◽  
pp. 39 ◽  
Author(s):  
Abbas Najim Salman ◽  
Maymona Ameen

<p>This paper is concerned with minimax shrinkage estimator using double stage shrinkage technique for lowering the mean squared error, intended for estimate the shape parameter (a) of Generalized Rayleigh distribution in a region (R) around available prior knowledge (a<sub>0</sub>) about the actual value (a) as initial estimate in case when the scale parameter (l) is known .</p><p>In situation where the experimentations are time consuming or very costly, a double stage procedure can be used to reduce the expected sample size needed to obtain the estimator.</p><p>The proposed estimator is shown to have smaller mean squared error for certain choice of the shrinkage weight factor y(<strong>×</strong>) and suitable region R.</p><p>Expressions for Bias, Mean squared error (MSE), Expected sample size [E (n/a, R)], Expected sample size proportion [E(n/a,R)/n], probability for avoiding the second sample and percentage of overall sample saved  for the proposed estimator are derived.</p><p>Numerical results and conclusions for the expressions mentioned above were displayed when the consider estimator are testimator of level of significanceD.</p><p>Comparisons with the minimax estimator and with the most recent studies were made to shown the effectiveness of the proposed estimator.</p>


2020 ◽  
Vol 2020 ◽  
pp. 1-22
Author(s):  
Byung-Kwon Son ◽  
Do-Jin An ◽  
Joon-Ho Lee

In this paper, a passive localization of the emitter using noisy angle-of-arrival (AOA) measurements, called Brown DWLS (Distance Weighted Least Squares) algorithm, is considered. The accuracy of AOA-based localization is quantified by the mean-squared error. Various estimates of the AOA-localization algorithm have been derived (Doğançay and Hmam, 2008). Explicit expression of the location estimate of the previous study is used to get an analytic expression of the mean-squared error (MSE) of one of the various estimates. To validate the derived expression, we compare the MSE from the Monte Carlo simulation with the analytically derived MSE.


2009 ◽  
Vol 106 (3) ◽  
pp. 975-983 ◽  
Author(s):  
Mark Burnley

To determine whether the asymptote of the torque-duration relationship (critical torque) could be estimated from the torque measured at the end of a series of maximal voluntary contractions (MVCs) of the quadriceps, eight healthy men performed eight laboratory tests. Following familiarization, subjects performed two tests in which they were required to perform 60 isometric MVCs over a period of 5 min (3 s contraction, 2 s rest), and five tests involving intermittent isometric contractions at ∼35–60% MVC, each performed to task failure. Critical torque was determined using linear regression of the torque impulse and contraction time during the submaximal tests, and the end-test torque during the MVCs was calculated from the mean of the last six contractions of the test. During the MVCs voluntary torque declined from 263.9 ± 44.6 to 77.8 ± 17.8 N·m. The end-test torque was not different from the critical torque (77.9 ± 15.9 N·m; 95% paired-sample confidence interval, −6.5 to 6.2 N·m). The root mean squared error of the estimation of critical torque from the end-test torque was 7.1 N·m. Twitch interpolation showed that voluntary activation declined from 90.9 ± 6.5% to 66.9 ± 13.1% ( P < 0.001), and the potentiated doublet response declined from 97.7 ± 23.0 to 46.9 ± 6.7 N·m ( P < 0.001) during the MVCs, indicating the development of both central and peripheral fatigue. These data indicate that fatigue during 5 min of intermittent isometric MVCs of the quadriceps leads to an end-test torque that closely approximates the critical torque.


Author(s):  
MOULOUD ADEL ◽  
DANIEL ZUWALA ◽  
MONIQUE RASIGNI ◽  
SALAH BOURENNANE

A noise reduction scheme on digitized mammographic phantom images is presented. This algorithm is based on a direct contrast modification method with an optimal function, obtained by using the mean squared error as a criterion. Computer simulated images containing objects similar to those observed in the phantom are built to evaluate the performance of the algorithm. Noise reduction results obtained on both simulated and real phantom images show that the developed method may be considered as a good preprocessing step from the point of view of automating phantom film evaluation by means of image processing.


2010 ◽  
Vol 1 (4) ◽  
pp. 17-45
Author(s):  
Antons Rebguns ◽  
Diana F. Spears ◽  
Richard Anderson-Sprecher ◽  
Aleksey Kletsov

This paper presents a novel theoretical framework for swarms of agents. Before deploying a swarm for a task, it is advantageous to predict whether a desired percentage of the swarm will succeed. The authors present a framework that uses a small group of expendable “scout” agents to predict the success probability of the entire swarm, thereby preventing many agent losses. The scouts apply one of two formulas to predict – the standard Bernoulli trials formula or the new Bayesian formula. For experimental evaluation, the framework is applied to simulated agents navigating around obstacles to reach a goal location. Extensive experimental results compare the mean-squared error of the predictions of both formulas with ground truth, under varying circumstances. Results indicate the accuracy and robustness of the Bayesian approach. The framework also yields an intriguing result, namely, that both formulas usually predict better in the presence of (Lennard-Jones) inter-agent forces than when their independence assumptions hold.


Sign in / Sign up

Export Citation Format

Share Document