scholarly journals Matching as Nonparametric Preprocessing for Reducing Model Dependence in Parametric Causal Inference

2007 ◽  
Vol 15 (3) ◽  
pp. 199-236 ◽  
Author(s):  
Daniel E. Ho ◽  
Kosuke Imai ◽  
Gary King ◽  
Elizabeth A. Stuart

Although published works rarely include causal estimates from more than a few model specifications, authors usually choose the presented estimates from numerous trial runs readers never see. Given the often large variation in estimates across choices of control variables, functional forms, and other modeling assumptions, how can researchers ensure that the few estimates presented are accurate or representative? How do readers know that publications are not merely demonstrations that it ispossibleto find a specification that fits the author's favorite hypothesis? And how do we evaluate or even define statistical properties like unbiasedness or mean squared error when no unique model or estimator even exists? Matching methods, which offer the promise of causal inference with fewer assumptions, constitute one possible way forward, but crucial results in this fast-growing methodological literature are often grossly misinterpreted. We explain how to avoid these misinterpretations and propose a unified approach that makes it possible for researchers to preprocess data with matching (such as with the easy-to-use software we offer) and then to apply the best parametric techniques they would have used anyway. This procedure makes parametric models produce more accurate and considerably less model-dependent causal inferences.

2021 ◽  
pp. 1-16
Author(s):  
Kevin Kloos

The use of machine learning algorithms at national statistical institutes has increased significantly over the past few years. Applications range from new imputation schemes to new statistical output based entirely on machine learning. The results are promising, but recent studies have shown that the use of machine learning in official statistics always introduces a bias, known as misclassification bias. Misclassification bias does not occur in traditional applications of machine learning and therefore it has received little attention in the academic literature. In earlier work, we have collected existing methods that are able to correct misclassification bias. We have compared their statistical properties, including bias, variance and mean squared error. In this paper, we present a new generic method to correct misclassification bias for time series and we derive its statistical properties. Moreover, we show numerically that it has a lower mean squared error than the existing alternatives in a wide variety of settings. We believe that our new method may improve machine learning applications in official statistics and we aspire that our work will stimulate further methodological research in this area.


Biometrika ◽  
2020 ◽  
Author(s):  
D C Ahfock ◽  
W J Astle ◽  
S Richardson

Summary Sketching is a probabilistic data compression technique that has been largely developed by the computer science community. Numerical operations on big datasets can be intolerably slow; sketching algorithms address this issue by generating a smaller surrogate dataset. Typically, inference proceeds on the compressed dataset. Sketching algorithms generally use random projections to compress the original dataset, and this stochastic generation process makes them amenable to statistical analysis. We argue that the sketched data can be modelled as a random sample, thus placing this family of data compression methods firmly within an inferential framework. In particular, we focus on the Gaussian, Hadamard and Clarkson–Woodruff sketches and their use in single-pass sketching algorithms for linear regression with huge samples. We explore the statistical properties of sketched regression algorithms and derive new distributional results for a large class of sketching estimators. A key result is a conditional central limit theorem for data-oblivious sketches. An important finding is that the best choice of sketching algorithm in terms of mean squared error is related to the signal-to-noise ratio in the source dataset. Finally, we demonstrate the theory and the limits of its applicability on two datasets.


2012 ◽  
Vol 20 (1) ◽  
pp. 1-24 ◽  
Author(s):  
Stefano M. Iacus ◽  
Gary King ◽  
Giuseppe Porro

We discuss a method for improving causal inferences called “Coarsened Exact Matching” (CEM), and the new “Monotonic Imbalance Bounding” (MIB) class of matching methods from which CEM is derived. We summarize what is known about CEM and MIB, derive and illustrate several new desirable statistical properties of CEM, and then propose a variety of useful extensions. We show that CEM possesses a wide range of statistical properties not available in most other matching methods but is at the same time exceptionally easy to comprehend and use. We focus on the connection between theoretical properties and practical applications. We also make available easy-to-use open source software forR, Stata, andSPSSthat implement all our suggestions.


Biometrika ◽  
2017 ◽  
Vol 104 (4) ◽  
pp. 845-861 ◽  
Author(s):  
Takamichi Baba ◽  
Takayuki Kanemori ◽  
Yoshiyuki Ninomiya

Summary For marginal structural models, which play an important role in causal inference, we consider a model selection problem within a semiparametric framework using inverse-probability-weighted estimation or doubly robust estimation. In this framework, the modelling target is a potential outcome that may be missing, so there is no classical information criterion. We define a mean squared error for treating the potential outcome and derive an asymptotic unbiased estimator as a $C_{p}$ criterion using an ignorable treatment assignment condition. Simulation shows that the proposed criterion outperforms a conventional one by providing smaller squared errors and higher frequencies of selecting the true model in all the settings considered. Moreover, in a real-data analysis we found a clear difference between the two criteria.


1991 ◽  
Vol 01 (01) ◽  
pp. 19-42 ◽  
Author(s):  
WASFY B. MIKHAEL ◽  
FRANK H. WU

In this paper, a unified approach for generating fast block- and sequential-gradient LMS FIR tapped delay line (TDL) adaptive algorithms is presented. These algorithms employ time-varying convergence factors which are tailored for the adaptive filter coefficients and updated at each block or single data iteration. The convergence factors are chosen to minimize the mean squared error (MSE) and are easily computed from readily available signals. The general formulation leads to three classes of adaptive algorithms. These algorithms. ordered in a descending order of their computational complexity and performance. are: the optimum block adaptive algorithm with individual adaptation of parameters (OBAI), the optimum block adaptive (OBA) and OBA shifting (ODAS) algorithms, and the homogeneous adaptive (HA) algorithm. In this paper, it is shown how each class of algorithms is obtained from the previous one, by a simple trade-off between adaptation performance and computational complexity. Implementation aspects of the generated algorithms are examined and their performance is evaluated and compared with several recently proposed algorithms by means of computer simulations under a wide range of adaptation conditions. The evaluation results show that the generated algorithms have attractive features in the comparisons due to the considerable reduction in the number of iterations required for a given adaptation accuracy. The improvement, however. is achieved at the expense of a relatively modest increase in the number of computations per data sample.


2012 ◽  
Vol 61 (2) ◽  
pp. 277-290 ◽  
Author(s):  
Ádám Csorba ◽  
Vince Láng ◽  
László Fenyvesi ◽  
Erika Michéli

Napjainkban egyre nagyobb igény mutatkozik olyan technológiák és módszerek kidolgozására és alkalmazására, melyek lehetővé teszik a gyors, költséghatékony és környezetbarát talajadat-felvételezést és kiértékelést. Ezeknek az igényeknek felel meg a reflektancia spektroszkópia, mely az elektromágneses spektrum látható (VIS) és közeli infravörös (NIR) tartományában (350–2500 nm) végzett reflektancia-mérésekre épül. Figyelembe véve, hogy a talajokról felvett reflektancia spektrum információban nagyon gazdag, és a vizsgált tartományban számos talajalkotó rendelkezik karakterisztikus spektrális „ujjlenyomattal”, egyetlen görbéből lehetővé válik nagyszámú, kulcsfontosságú talajparaméter egyidejű meghatározása. Dolgozatunkban, a reflektancia spektroszkópia alapjaira helyezett, a talajok ösz-szetételének meghatározását célzó módszertani fejlesztés első lépéseit mutatjuk be. Munkánk során talajok szervesszén- és CaCO3-tartalmának megbecslését lehetővé tévő többváltozós matematikai-statisztikai módszerekre (részleges legkisebb négyzetek módszere, partial least squares regression – PLSR) épülő prediktív modellek létrehozását és tesztelését végeztük el. A létrehozott modellek tesztelése során megállapítottuk, hogy az eljárás mindkét talajparaméter esetében magas R2értéket [R2(szerves szén) = 0,815; R2(CaCO3) = 0,907] adott. A becslés pontosságát jelző közepes négyzetes eltérés (root mean squared error – RMSE) érték mindkét paraméter esetében közepesnek mondható [RMSE (szerves szén) = 0,467; RMSE (CaCO3) = 3,508], mely a reflektancia mérési előírások standardizálásával jelentősen javítható. Vizsgálataink alapján arra a következtetésre jutottunk, hogy a reflektancia spektroszkópia és a többváltozós kemometriai eljárások együttes alkalmazásával, gyors és költséghatékony adatfelvételezési és -értékelési módszerhez juthatunk.


Author(s):  
Nadia Hashim Al-Noor ◽  
Shurooq A.K. Al-Sultany

        In real situations all observations and measurements are not exact numbers but more or less non-exact, also called fuzzy. So, in this paper, we use approximate non-Bayesian computational methods to estimate inverse Weibull parameters and reliability function with fuzzy data. The maximum likelihood and moment estimations are obtained as non-Bayesian estimation. The maximum likelihood estimators have been derived numerically based on two iterative techniques namely “Newton-Raphson” and the “Expectation-Maximization” techniques. In addition, we provide compared numerically through Monte-Carlo simulation study to obtained estimates of the parameters and reliability function in terms of their mean squared error values and integrated mean squared error values respectively.


2014 ◽  
Vol 2 (2) ◽  
pp. 47-58
Author(s):  
Ismail Sh. Baqer

A two Level Image Quality enhancement is proposed in this paper. In the first level, Dualistic Sub-Image Histogram Equalization DSIHE method decomposes the original image into two sub-images based on median of original images. The second level deals with spikes shaped noise that may appear in the image after processing. We presents three methods of image enhancement GHE, LHE and proposed DSIHE that improve the visual quality of images. A comparative calculations is being carried out on above mentioned techniques to examine objective and subjective image quality parameters e.g. Peak Signal-to-Noise Ratio PSNR values, entropy H and mean squared error MSE to measure the quality of gray scale enhanced images. For handling gray-level images, convenient Histogram Equalization methods e.g. GHE and LHE tend to change the mean brightness of an image to middle level of the gray-level range limiting their appropriateness for contrast enhancement in consumer electronics such as TV monitors. The DSIHE methods seem to overcome this disadvantage as they tend to preserve both, the brightness and contrast enhancement. Experimental results show that the proposed technique gives better results in terms of Discrete Entropy, Signal to Noise ratio and Mean Squared Error values than the Global and Local histogram-based equalization methods


Geosciences ◽  
2020 ◽  
Vol 10 (9) ◽  
pp. 329
Author(s):  
Mahdi O. Karkush ◽  
Mahmood D. Ahmed ◽  
Ammar Abdul-Hassan Sheikha ◽  
Ayad Al-Rumaithi

The current study involves placing 135 boreholes drilled to a depth of 10 m below the existing ground level. Three standard penetration tests (SPT) are performed at depths of 1.5, 6, and 9.5 m for each borehole. To produce thematic maps with coordinates and depths for the bearing capacity variation of the soil, a numerical analysis was conducted using MATLAB software. Despite several-order interpolation polynomials being used to estimate the bearing capacity of soil, the first-order polynomial was the best among the other trials due to its simplicity and fast calculations. Additionally, the root mean squared error (RMSE) was almost the same for the all of the tried models. The results of the study can be summarized by the production of thematic maps showing the variation of the bearing capacity of the soil over the whole area of Al-Basrah city correlated with several depths. The bearing capacity of soil obtained from the suggested first-order polynomial matches well with those calculated from the results of SPTs with a deviation of ±30% at a 95% confidence interval.


Sign in / Sign up

Export Citation Format

Share Document