msreg: A command for consistent estimation of linear regression models using matched data

Author(s):  
Masayuki Hirukawa ◽  
Di Liu ◽  
Artem Prokhorov

Economists often use matched samples, especially when dealing with earning data where some observations are missing in one sample and need to be imputed from another sample. Hirukawa and Prokhorov (2018, Journal of Econometrics 203: 344–358) show that the ordinary least-squares estimator using matched samples is inconsistent and propose two consistent estimators. We describe a new command, msreg, that implements these two consistent estimators based on two samples. The estimators attain the parametric convergence rate if the number of continuous matching variables is no greater than four.

Author(s):  
Warha, Abdulhamid Audu ◽  
Yusuf Abbakar Muhammad ◽  
Akeyede, Imam

Linear regression is the measure of relationship between two or more variables known as dependent and independent variables. Classical least squares method for estimating regression models consist of minimising the sum of the squared residuals. Among the assumptions of Ordinary least squares method (OLS) is that there is no correlations (multicollinearity) between the independent variables. Violation of this assumptions arises most often in regression analysis and can lead to inefficiency of the least square method. This study, therefore, determined the efficient estimator between Least Absolute Deviation (LAD) and Weighted Least Square (WLS) in multiple linear regression models at different levels of multicollinearity in the explanatory variables. Simulation techniques were conducted using R Statistical software, to investigate the performance of the two estimators under violation of assumptions of lack of multicollinearity. Their performances were compared at different sample sizes. Finite properties of estimators’ criteria namely, mean absolute error, absolute bias and mean squared error were used for comparing the methods. The best estimator was selected based on minimum value of these criteria at a specified level of multicollinearity and sample size. The results showed that, LAD was the best at different levels of multicollinearity and was recommended as alternative to OLS under this condition. The performances of the two estimators decreased when the levels of multicollinearity was increased.


Author(s):  
Aditio Putra G ◽  
Muhammad Arif Tiro ◽  
Muhammad Kasim Aidid

Abstrak Metode kuadrat terkecil merupakan metode standar untuk mengestimasi nilai parameter model regresi linear. Metode tersebut dibangun berdasarkan asumsi error bersifat identik dan independen, serta berdistribusi normal. Apabila asumsi tidak terpenuhi maka metode ini tidak akurat. Alternatif untuk mengatasi hal tersebut adalah dengan menggunakan metode resampling. Adapun metode resampling yang digunakan dalam penelitian ini yaitu metode bootstrap dan Jackknife. Terlebih dahulu dilakukan estimasi nilai parameter regresi untuk analisis data kemiskinan Kota Makassar Tahun 2017. Data tersebut merupakan data sekunder diperoleh dari BAPPEDA Kota Makassar. Dari uji asumsi klasik diperoleh bahwa model tidak bersifat homoskedastis dan residual tidak berdistribusi normal sehingga model regresi yang diperoleh tidak dapat dipertanggungjawabkan. Metode bootstrap dan jackknife yang dikenalkan disini menggunakan program R untuk mencari nilai bias dan nilai standar errornya. Estimasi parameter model regresi linear berganda dari metode resampling bootstrap dengan B=200 dan B=500 serta metode resampling jackknife Terhapus-1 diperoleh model regresi. Hasil yang didapat dalam penelitian ini, metode jackknife merupakan metode yang efisien dibandingkan dengan metode bootstrap, hal ini didukung dengan kecilnya tingkat standar error dan nilai biasnya yang dihasilkan. Kata Kunci: Regrei, Resampling, Bootsrap, JaccknifeAbstract. The Ordinary least squares method is a standard method for estimating the parameter values of a linear regression model. The method is built based on error assumptions that are identical and independent, and are normally distributed. If the assumptions are not met, this method is not accurate. The alternative to overcome this is to use the resampling method. The resampling method used in this study is bootstrap and jackknife methods. First, estimation of regression parameter values for analysis of poverty data in Makassar City in 2017. The data is secondary data obtained from the BAPPEDA of Makassar City. From the classic assumption test, it is obtained that the model is not homosexedastic and residual is not normally distributed so that the regression model obtained cannot be accounted for. Bootstrap and jackknife methods are introduced here using the R program to find the value of the bias and the standard error values. Parameter estimation of multiple linear regression models from Bootstrap resampling method with B= 200, B= 500 and jackknife deleted-1 resampling method obtained regression models. The results obtained in this study, Jackknife method is an efficient method compared with the bootstrap method, and this is supported by the small standard level error and bias in resulting value.Keywords: regression, resampling, bootstrap, jackknife.


Stats ◽  
2020 ◽  
Vol 3 (4) ◽  
pp. 526-541
Author(s):  
Issam Dawoud ◽  
B. M. Golam Kibria

In a multiple linear regression model, the ordinary least squares estimator is inefficient when the multicollinearity problem exists. Many authors have proposed different estimators to overcome the multicollinearity problem for linear regression models. This paper introduces a new regression estimator, called the Dawoud–Kibria estimator, as an alternative to the ordinary least squares estimator. Theory and simulation results show that this estimator performs better than other regression estimators under some conditions, according to the mean squares error criterion. The real-life datasets are used to illustrate the findings of the paper.


Author(s):  
Maria Boilé ◽  
Michail Golias

This paper employs linear regression algorithms in order to train models under the presence of limited training data. Usually in transportation applications, these models are built via Ordinary Least Squares and Stepwise Regression, which perform poorly under limited data. The algorithms presented in this paper have been extensively used in other scientific fields for problems with similar conditions and seem to partially or fully remedy this problem and its consequences. Four different algorithms are presented and several models are built. The models are used for truck volume prediction on highway sections in New Jersey, and results are compared to Stepwise Linear regression models.


2018 ◽  
Vol 23 (1) ◽  
pp. 60-71
Author(s):  
Wigiyanti Masodah

Offering credit is the main activity of a Bank. There are some considerations when a bank offers credit, that includes Interest Rates, Inflation, and NPL. This study aims to find out the impact of Variable Interest Rates, Inflation variables and NPL variables on credit disbursed. The object in this study is state-owned banks. The method of analysis in this study uses multiple linear regression models. The results of the study have shown that Interest Rates and NPL gave some negative impacts on the given credit. Meanwhile, Inflation variable does not have a significant effect on credit given. Keywords: Interest Rate, Inflation, NPL, offered Credit.


Sign in / Sign up

Export Citation Format

Share Document