empirical risk
Recently Published Documents


TOTAL DOCUMENTS

244
(FIVE YEARS 85)

H-INDEX

19
(FIVE YEARS 2)

2022 ◽  
pp. 1-20
Author(s):  
Salim Moudache ◽  
◽  
Mourad Badri

This work aims to investigate the potential, from different perspectives, of a risk model to support Cross-Version Fault and Severity Prediction (CVFSP) in object-oriented software. The risk of a class is addressed from the perspective of two particular factors: the number of faults it can contain and their severity. We used various object-oriented metrics to capture the two risk factors. The risk of a class is modeled using the concept of Euclidean distance. We used a dataset collected from five successive versions of an open-source Java software system (ANT). We investigated different variants of the considered risk model, based on various combinations of object-oriented metrics pairs. We used different machine learning algorithms for building the prediction models: Naive Bayes (NB), J48, Random Forest (RF), Support Vector Machines (SVM) and Multilayer Perceptron (ANN). We investigated the effectiveness of the prediction models for Cross-Version Fault and Severity Prediction (CVFSP), using data of prior versions of the considered system. We also investigated if the considered risk model can give as output the Empirical Risk (ER) of a class, a continuous value considering both the number of faults and their different levels of severity. We used different techniques for building the prediction models: Linear Regression (LR), Gaussian Process (GP), Random forest (RF) and M5P (two decision trees algorithms), SmoReg and Artificial Neural Network (ANN). The considered risk model achieves acceptable results for both cross-version binary fault prediction (a g-mean of 0.714, an AUC of 0.725) and cross-version multi-classification of levels of severity (a g-mean of 0.758, an AUC of 0.771). The model also achieves good results in the estimation of the empirical risk of a class by considering both the number of faults and their levels of severity (intra-version analysis with a correlation coefficient of 0.659, cross-version analysis with a correlation coefficient of 0.486).


2021 ◽  
Vol 2021 (12) ◽  
pp. 124015
Author(s):  
Fabrizio Pittorino ◽  
Carlo Lucibello ◽  
Christoph Feinauer ◽  
Gabriele Perugini ◽  
Carlo Baldassi ◽  
...  

Abstract The properties of flat minima in the empirical risk landscape of neural networks have been debated for some time. Increasing evidence suggests they possess better generalization capabilities with respect to sharp ones. In this work we first discuss the relationship between alternative measures of flatness: the local entropy, which is useful for analysis and algorithm development, and the local energy, which is easier to compute and was shown empirically in extensive tests on state-of-the-art networks to be the best predictor of generalization capabilities. We show semi-analytically in simple controlled scenarios that these two measures correlate strongly with each other and with generalization. Then, we extend the analysis to the deep learning scenario by extensive numerical validations. We study two algorithms, entropy-stochastic gradient descent and replicated-stochastic gradient descent, that explicitly include the local entropy in the optimization objective. We devise a training schedule by which we consistently find flatter minima (using both flatness measures), and improve the generalization error for common architectures (e.g. ResNet, EfficientNet).


2021 ◽  
pp. 174569162110178
Author(s):  
Erika A. Waters ◽  
Marc T. Kiviniemi ◽  
Jennifer L. Hay ◽  
Heather Orom

Since the middle of the 20th century, perceptions of risk have been critical to understanding engagement in volitional behavior change. However, theoretical and empirical risk perception research seldom considers the possibility that risk perceptions do not simply exist: They must be formed. Thus, some people may not have formulated a perception of risk for a hazard at the time a researcher asks them, or they may not be confident in the extent to which their perception matches reality. We describe a decade-long research program that investigates the possibility that some people may genuinely not know their risk of even well-publicized hazards. We demonstrate that indications of not knowing (i.e., “don’t know” responses) are prevalent in the U.S. population, are systematically more likely to occur among marginalized sociodemographic groups, and are associated with less engagement in protective health behaviors. “Don’t know” responses are likely indications of genuinely limited knowledge and therefore may indicate populations in need of targeted intervention. This body of research suggests that not allowing participants to indicate their uncertainty may threaten the validity and generalizability of behavior-change research. We provide concrete recommendations for scientists to allow participants to express uncertainty and to analyze the resulting data.


Author(s):  
Zhengling Qi ◽  
Ying Cui ◽  
Yufeng Liu ◽  
Jong-Shi Pang

This paper has two main goals: (a) establish several statistical properties—consistency, asymptotic distributions, and convergence rates—of stationary solutions and values of a class of coupled nonconvex and nonsmooth empirical risk-minimization problems and (b) validate these properties by a noisy amplitude-based phase-retrieval problem, the latter being of much topical interest. Derived from available data via sampling, these empirical risk-minimization problems are the computational workhorse of a population risk model that involves the minimization of an expected value of a random functional. When these minimization problems are nonconvex, the computation of their globally optimal solutions is elusive. Together with the fact that the expectation operator cannot be evaluated for general probability distributions, it becomes necessary to justify whether the stationary solutions of the empirical problems are practical approximations of the stationary solution of the population problem. When these two features, general distribution and nonconvexity, are coupled with nondifferentiability that often renders the problems “non-Clarke regular,” the task of the justification becomes challenging. Our work aims to address such a challenge within an algorithm-free setting. The resulting analysis is, therefore, different from much of the analysis in the recent literature that is based on local search algorithms. Furthermore, supplementing the classical global minimizer-centric analysis, our results offer a promising step to close the gap between computational optimization and asymptotic analysis of coupled, nonconvex, nonsmooth statistical estimation problems, expanding the former with statistical properties of the practically obtained solution and providing the latter with a more practical focus pertaining to computational tractability.


2021 ◽  
pp. 1-42
Author(s):  
Ilsang Ohn ◽  
Yongdai Kim

Abstract Recent theoretical studies proved that deep neural network (DNN) estimators obtained by minimizing empirical risk with a certain sparsity constraint can attain optimal convergence rates for regression and classification problems. However, the sparsity constraint requires knowing certain properties of the true model, which are not available in practice. Moreover, computation is difficult due to the discrete nature of the sparsity constraint. In this letter, we propose a novel penalized estimation method for sparse DNNs that resolves the problems existing in the sparsity constraint. We establish an oracle inequality for the excess risk of the proposed sparse-penalized DNN estimator and derive convergence rates for several learning tasks. In particular, we prove that the sparse-penalized estimator can adaptively attain minimax convergence rates for various nonparametric regression problems. For computation, we develop an efficient gradient-based optimization algorithm that guarantees the monotonic reduction of the objective function.


Author(s):  
Sandeep Samantaray ◽  
Dillip Kumar Ghose

Abstract Effective prediction of runoff is a substantial feature for the successful management of hydrological phenomena in arid regions. The present research findings reveal that a rainfall simulator (RS) can be a valuable instrument to estimate runoff as the intensity of rainfall is modifiable in the course of an experimental process, which turns out to be of great advantage. Rainfall–runoff process is a complex physical phenomenon caused by the effect of various parameters. In this research, a new hybrid technique integrating PSR (phase space reconstruction) with FFA (firefly algorithm) and SVM (support vector machine) has gained recognition in various modelling investigations in contrast to the principle of empirical risk minimization through ANN practices. Outcomes of SVM are contrasted against SVM-FFA and PSR-SVM-FFA models. The improvements in NSE (Nash–Sutcliffe), RMSE (Root Mean Square Error), and WI (Willmott's Index) by PSR-SVM-FFA over SVM models specify that the prediction accuracy of the hybrid model is better. The established PSR-SVM-FFA model generates preeminent WI values that range from 0.97 to 0.98, while the SVM and SVM-FFA models encompass 0.93–0.95 and 0.96–0.97, respectively. The proposed PSR-SVM-FFA model gives more accurate results and error limiting up to 2–3%.


2021 ◽  
pp. 1-52
Author(s):  
Taira Tsuchiya ◽  
Nontawat Charoenphakdee ◽  
Issei Sato ◽  
Masashi Sugiyama

Abstract Ordinal regression is aimed at predicting an ordinal class label. In this letter, we consider its semisupervised formulation, in which we have unlabeled data along with ordinal-labeled data to train an ordinal regressor. There are several metrics to evaluate the performance of ordinal regression, such as the mean absolute error, mean zero-one error, and mean squared error. However, the existing studies do not take the evaluation metric into account, restrict model choice, and have no theoretical guarantee. To overcome these problems, we propose a novel generic framework for semisupervised ordinal regression based on the empirical risk minimization principle that is applicable to optimizing all of the metrics mentioned above. In addition, our framework has flexible choices of models, surrogate losses, and optimization algorithms without the common geometric assumption on unlabeled data such as the cluster assumption or manifold assumption. We provide an estimation error bound to show that our risk estimator is consistent. Finally, we conduct experiments to show the usefulness of our framework.


Entropy ◽  
2021 ◽  
Vol 23 (10) ◽  
pp. 1255
Author(s):  
Yuheng Bu ◽  
Weihao Gao ◽  
Shaofeng Zou ◽  
Venugopal V. Veeravalli

It has been reported in many recent works on deep model compression that the population risk of a compressed model can be even better than that of the original model. In this paper, an information-theoretic explanation for this population risk improvement phenomenon is provided by jointly studying the decrease in the generalization error and the increase in the empirical risk that results from model compression. It is first shown that model compression reduces an information-theoretic bound on the generalization error, which suggests that model compression can be interpreted as a regularization technique to avoid overfitting. The increase in empirical risk caused by model compression is then characterized using rate distortion theory. These results imply that the overall population risk could be improved by model compression if the decrease in generalization error exceeds the increase in empirical risk. A linear regression example is presented to demonstrate that such a decrease in population risk due to model compression is indeed possible. Our theoretical results further suggest a way to improve a widely used model compression algorithm, i.e., Hessian-weighted K-means clustering, by regularizing the distance between the clustering centers. Experiments with neural networks are provided to validate our theoretical assertions.


Sign in / Sign up

Export Citation Format

Share Document