scholarly journals Discussion of Identification, Estimation and Approximation of Risk under Interventions that Depend on the Natural Value of Treatment Using Observational Data, by Jessica Young, Miguel Hernán, and James Robins

2014 ◽  
Vol 3 (1) ◽  
Author(s):  
Mark J. van der Laan ◽  
Alexander R. Luedtke ◽  
Iván Díaz

AbstractYoung, Hernán, and Robins consider the mean outcome under a dynamic intervention that may rely on the natural value of treatment. They first identify this value with a statistical target parameter, and then show that this statistical target parameter can also be identified with a causal parameter which gives the mean outcome under a stochastic intervention. The authors then describe estimation strategies for these quantities. Here we augment the authors’ insightful discussion by sharing our experiences in situations where two causal questions lead to the same statistical estimand, or the newer problem that arises in the study of data adaptive parameters, where two statistical estimands can lead to the same estimation problem. Given a statistical estimation problem, we encourage others to always use a robust estimation framework where the data generating distribution truly belongs to the statistical model. We close with a discussion of a framework which has these properties.

2013 ◽  
Vol 4 (3) ◽  
pp. 32-53
Author(s):  
Peter Schanbacher

Many social interactions (examples are market overreactions, high rates of acquisitions, strikes, wars) are the result of agents' overconfidence. Agents are in particular overconfident for difficult tasks. This paper analyzes overconfidence in the context of a statistical estimation problem. The authors find that it is rational to (i) be overconfident and (ii) to be notably overconfident if the task is difficult. The counterintuitive finding that uninformed agents which should be the least confident ones show the highest degree of overconfidence can be explained as a rational behavior.


2019 ◽  
Vol 489 (1) ◽  
pp. 842-854 ◽  
Author(s):  
Dandan Xu ◽  
Ling Zhu ◽  
Robert Grand ◽  
Volker Springel ◽  
Shude Mao ◽  
...  

ABSTRACT Motivated by the recently discovered kinematic ‘Hubble sequence’ shown by the stellar orbit-circularity distribution of 260 CALIFA galaxies, we make use of a comparable galaxy sample at z = 0 with a stellar mass range of $M_{*}/\mathrm{M}_{\odot }\in [10^{9.7},\, 10^{11.4}]$ selected from the IllustrisTNG simulation and study their stellar orbit compositions in relation to a number of other fundamental galaxy properties. We find that the TNG100 simulation broadly reproduces the observed fractions of different orbital components and their stellar mass dependences. In particular, the mean mass dependences of the luminosity fractions for the kinematically warm and hot orbits are well reproduced within model uncertainties of the observed galaxies. The simulation also largely reproduces the observed peak and trough features at $M_{*}\approx 1\rm {-}2\times 10^{10}\, \mathrm{M}_{\odot }$ in the mean distributions of the cold- and hot-orbit fractions, respectively, indicating fewer cooler orbits and more hotter orbits in both more- and less-massive galaxies beyond such a mass range. Several marginal disagreements are seen between the simulation and observations: the average cold-orbit (counter-rotating) fractions of the simulated galaxies below (above) $M_{*}\approx 6\times 10^{10}\, \mathrm{M}_{\odot }$ are systematically higher than the observational data by $\lesssim 10{{\ \rm per\ cent}}$ (absolute orbital fraction); the simulation also seems to produce more scatter for the cold-orbit fraction and less so for the non-cold orbits at any given galaxy mass. Possible causes that stem from the adopted heating mechanisms are discussed.


Author(s):  
M. H Badii

Keywords: Estimations, sampling, statisticsAbstract. The notion of statistical estimation both in terms of point and interval is described. The criteria of a good estimator are noted. The procedures to calculate the intervals for the mean, proportions and the difference among two means as well as the confidence intervals for the probable errors in statistics are provided.Palabras clave: Estadística, estimación, muestreoResumen. En la presente investigación se describen la noción de la estimación estadística, tanto de tipo puntual con de forma de intervalo. Se presentan los criterios que debe reunir un estimador bueno. Se notan con ejemplos, la forma de calcular la estimación del intervalo para la media, la proporción y de la diferencia entre dos medias y los intervalos de confianza para los errores probables.


2003 ◽  
Vol 125 (3) ◽  
pp. 791-796 ◽  
Author(s):  
D. Taraza

The goal of this two-part paper is to develop a methodology using the variation of the measured crankshaft speed to calculate the mean indicated pressure (MIP) of a multicylinder engine and to detect cylinders that are lower contributors to the total engine output. Both the gas pressure torque and the crankshaft’s speed are, under steady-state operating conditions, periodic functions of the crank angle and may be expressed by Fourier series. For the lower harmonic orders, the dynamic response of the crankshaft approaches the response of a rigid body and that makes it is possible to establish correlations between the amplitudes and phases of the corresponding harmonic orders of the crankshaft’s speed and of the gas pressure torque. The inherent cycle-to-cycle variation in the operation of the cylinders requires a statistical approach to the problem. The first part of the paper introduces the statistical model for a harmonic component of the gas pressure torque and determines the correlation between the amplitudes and phases of the harmonic components of the gas pressure torque and the MIP of the engine. In the second part of the paper the statistical model is used to calculate the MIP and to detect deficient cylinders in the operation of a six-cylinder four-stroke diesel engine.


2014 ◽  
Vol 2 (1) ◽  
pp. 13-74 ◽  
Author(s):  
Mark J. van der Laan

AbstractSuppose that we observe a population of causally connected units. On each unit at each time-point on a grid we observe a set of other units the unit is potentially connected with, and a unit-specific longitudinal data structure consisting of baseline and time-dependent covariates, a time-dependent treatment, and a final outcome of interest. The target quantity of interest is defined as the mean outcome for this group of units if the exposures of the units would be probabilistically assigned according to a known specified mechanism, where the latter is called a stochastic intervention. Causal effects of interest are defined as contrasts of the mean of the unit-specific outcomes under different stochastic interventions one wishes to evaluate. This covers a large range of estimation problems from independent units, independent clusters of units, and a single cluster of units in which each unit has a limited number of connections to other units. The allowed dependence includes treatment allocation in response to data on multiple units and so called causal interference as special cases. We present a few motivating classes of examples, propose a structural causal model, define the desired causal quantities, address the identification of these quantities from the observed data, and define maximum likelihood based estimators based on cross-validation. In particular, we present maximum likelihood based super-learning for this network data. Nonetheless, such smoothed/regularized maximum likelihood estimators are not targeted and will thereby be overly bias w.r.t. the target parameter, and, as a consequence, generally not result in asymptotically normally distributed estimators of the statistical target parameter.To formally develop estimation theory, we focus on the simpler case in which the longitudinal data structure is a point-treatment data structure. We formulate a novel targeted maximum likelihood estimator of this estimand and show that the double robustness of the efficient influence curve implies that the bias of the targeted minimum loss-based estimation (TMLE) will be a second-order term involving squared differences of two nuisance parameters. In particular, the TMLE will be consistent if either one of these nuisance parameters is consistently estimated. Due to the causal dependencies between units, the data set may correspond with the realization of a single experiment, so that establishing a (e.g. normal) limit distribution for the targeted maximum likelihood estimators, and corresponding statistical inference, is a challenging topic. We prove two formal theorems establishing the asymptotic normality using advances in weak-convergence theory. We conclude with a discussion and refer to an accompanying technical report for extensions to general longitudinal data structures.


2005 ◽  
Vol 35 (4) ◽  
pp. 444-457 ◽  
Author(s):  
Jeff A. Polton ◽  
David M. Lewis ◽  
Stephen E. Belcher

Abstract The interaction between the Coriolis force and the Stokes drift associated with ocean surface waves leads to a vertical transport of momentum, which can be expressed as a force on the mean momentum equation in the direction along wave crests. How this Coriolis–Stokes forcing affects the mean current profile in a wind-driven mixed layer is investigated using simple models, results from large-eddy simulations, and observational data. The effects of the Coriolis–Stokes forcing on the mean current profile are examined by reappraising analytical solutions to the Ekman model that include the Coriolis–Stokes forcing. Turbulent momentum transfer is modeled using an eddy-viscosity model, first with a constant viscosity and second with a linearly varying eddy viscosity. Although the Coriolis–Stokes forcing penetrates only a small fraction of the depth of the wind-driven layer for parameter values typical of the ocean, the analytical solutions show how the current profile is substantially changed through the whole depth of the wind-driven layer. It is shown how, for this oceanic regime, the Coriolis–Stokes forcing supports a fraction of the applied wind stress, changing the boundary condition on the wind-driven component of the flow and hence changing the current profile through all depths. The analytical solution with the linearly varying eddy viscosity is shown to reproduce reasonably well the effects of the Coriolis–Stokes forcing on the current profile computed from large-eddy simulations, which resolve the three-dimensional overturning motions associated with the turbulent Langmuir circulations in the wind-driven layer. Last, the analytical solution with the Coriolis–Stokes forcing is shown to agree reasonably well with current profiles from previously published observational data and certainly agrees better than the standard Ekman model. This finding provides evidence that the Coriolis–Stokes forcing is an important mechanism in controlling the dynamics of the upper ocean.


Sign in / Sign up

Export Citation Format

Share Document