The use of the Maximum Likelihood Estimator and the Dynamically Penalized Likelihood methods in inverse radiation therapy planning

Author(s):  
Jorge Llacer ◽  
Timothy D. Solberg ◽  
Claus Promberger ◽  
Alexander Kuzmany
Biometrika ◽  
2020 ◽  
Author(s):  
Ioannis Kosmidis ◽  
David Firth

Summary Penalization of the likelihood by Jeffreys’ invariant prior, or a positive power thereof, is shown to produce finite-valued maximum penalized likelihood estimates in a broad class of binomial generalized linear models. The class of models includes logistic regression, where the Jeffreys-prior penalty is known additionally to reduce the asymptotic bias of the maximum likelihood estimator, and models with other commonly used link functions, such as probit and log-log. Shrinkage towards equiprobability across observations, relative to the maximum likelihood estimator, is established theoretically and studied through illustrative examples. Some implications of finiteness and shrinkage for inference are discussed, particularly when inference is based on Wald-type procedures. A widely applicable procedure is developed for computation of maximum penalized likelihood estimates, by using repeated maximum likelihood fits with iteratively adjusted binomial responses and totals. These theoretical results and methods underpin the increasingly widespread use of reduced-bias and similarly penalized binomial regression models in many applied fields.


Author(s):  
Hazim Mansour Gorgees ◽  
Bushra Abdualrasool Ali ◽  
Raghad Ibrahim Kathum

     In this paper, the maximum likelihood estimator and the Bayes estimator of the reliability function for negative exponential distribution has been derived, then a Monte –Carlo simulation technique was employed to compare the performance of such estimators. The integral mean square error (IMSE) was used as a criterion for this comparison. The simulation results displayed that the Bayes estimator performed better than the maximum likelihood estimator for different samples sizes.


2021 ◽  
Author(s):  
Jakob Raymaekers ◽  
Peter J. Rousseeuw

AbstractMany real data sets contain numerical features (variables) whose distribution is far from normal (Gaussian). Instead, their distribution is often skewed. In order to handle such data it is customary to preprocess the variables to make them more normal. The Box–Cox and Yeo–Johnson transformations are well-known tools for this. However, the standard maximum likelihood estimator of their transformation parameter is highly sensitive to outliers, and will often try to move outliers inward at the expense of the normality of the central part of the data. We propose a modification of these transformations as well as an estimator of the transformation parameter that is robust to outliers, so the transformed data can be approximately normal in the center and a few outliers may deviate from it. It compares favorably to existing techniques in an extensive simulation study and on real data.


2013 ◽  
Vol 55 (3) ◽  
pp. 643-652
Author(s):  
Gauss M. Cordeiro ◽  
Denise A. Botter ◽  
Alexsandro B. Cavalcanti ◽  
Lúcia P. Barroso

2020 ◽  
Vol 28 (3) ◽  
pp. 183-196
Author(s):  
Kouacou Tanoh ◽  
Modeste N’zi ◽  
Armel Fabrice Yodé

AbstractWe are interested in bounds on the large deviations probability and Berry–Esseen type inequalities for maximum likelihood estimator and Bayes estimator of the parameter appearing linearly in the drift of nonhomogeneous stochastic differential equation driven by fractional Brownian motion.


2014 ◽  
Vol 42 (1) ◽  
pp. 391-399 ◽  
Author(s):  
Alexandra R. Cunliffe ◽  
Clay Contee ◽  
Samuel G. Armato ◽  
Bradley White ◽  
Julia Justusson ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document