Why Differential Data Work

Author(s):  
William Menke ◽  
Roger Creel

ABSTRACT This article explains the features of differential data that make them attractive, their shortcomings, and the situations for which they are best suited. The use of differential data is ubiquitous in the seismological community, in which they are used to determine earthquake locations via the double-difference method and the Earth’s velocity structure via geotomography; furthermore, they have important applications in other areas of geophysics, as well. A common assumption is that differential data are uncorrelated and have uniform variance. We show that this assumption is well justified when the original, undifferenced data covary with each other according to a two-sided exponential function. It is not well justified when they covary according to a Gaussian function. Differences of exponentially correlated data are approximately uncorrelated with uniform variance when they are regularly spaced in distance. However, when they are irregularly spaced, they are uncorrelated with a nonuniform variance that scales with the spacing of the data. When differential data are computed by taking differences of the original, undifferenced data, model parameters estimated using ordinary least squares applied to the differential data are almost exactly equal to those estimated using weighed least squares applied to the original, undifferenced data (with the weights given by the inverse covariance matrix). A better solution only results when the differential data are directly estimated and their variance is smaller than is implied by differencing the original data. Differential data may be appropriate for global seismic travel-time data because the covariance of errors in predicted travel times may have a covariance close to a two-sided exponential, on account of the upper mantle being close to a Von Karman medium with exponent κ≪12.

2020 ◽  
pp. 636-645
Author(s):  
Hussain Karim Nashoor ◽  
Ebtisam Karim Abdulah

Examination of skewness makes academics more aware of the importance of accurate statistical analysis. Undoubtedly, most phenomena contain a certain percentage of skewness which resulted to the appearance of what is -called "asymmetry" and, consequently, the importance of the skew normal family . The epsilon skew normal distribution ESN (μ, σ, ε) is one of the probability distributions which provide a more flexible model because the skewness parameter provides the possibility to fluctuate from normal to skewed distribution. Theoretically, the estimation of linear regression model parameters, with an average error value that is not zero, is considered a major challenge due to having difficulties, as no explicit formula to calculate these estimates can be obtained. Practically, values for these estimates can be obtained only by referring to numerical methods. This research paper is dedicated to estimate parameters of the Epsilon Skew Normal General Linear Model (ESNGLM) using an adaptive least squares method, as along with the employment of the ordinary least squares method for estimating parameters of the General Linear Model (GLM). In addition, the coefficient of determination was used as a criterion to compare the models’ preference. These methods were applied to real data represented by dollar exchange rates. The Matlab software was applied in this work and the results showed that the ESNGLM represents a satisfactory model. 


2014 ◽  
Vol 71 (1) ◽  
Author(s):  
Bello Abdulkadir Rasheed ◽  
Robiah Adnan ◽  
Seyed Ehsan Saffari ◽  
Kafi Dano Pati

In a linear regression model, the ordinary least squares (OLS) method is considered the best method to estimate the regression parameters if the assumptions are met. However, if the data does not satisfy the underlying assumptions, the results will be misleading. The violation for the assumption of constant variance in the least squares regression is caused by the presence of outliers and heteroscedasticity in the data. This assumption of constant variance (homoscedasticity) is very important in linear regression in which the least squares estimators enjoy the property of minimum variance. Therefor e robust regression method is required to handle the problem of outlier in the data. However, this research will use the weighted least square techniques to estimate the parameter of regression coefficients when the assumption of error variance is violated in the data. Estimation of WLS is the same as carrying out the OLS in a transformed variables procedure. The WLS can easily be affected by outliers. To remedy this, We have suggested a strong technique for the estimation of regression parameters in the existence of heteroscedasticity and outliers. Here we apply the robust regression of M-estimation using iterative reweighted least squares (IRWLS) of Huber and Tukey Bisquare function and resistance regression estimator of least trimmed squares to estimating the model parameters of state-wide crime of united states in 1993. The outcomes from the study indicate the estimators obtained from the M-estimation techniques and the least trimmed method are more effective compared with those obtained from the OLS.


Author(s):  
Anatolii Omelchenko ◽  
Oleksandr Vinnichenko ◽  
Pavel Neyezhmakov ◽  
Oleksii Fedorov ◽  
Volodymyr Bolyuh

Abstract In order to develop optimal data processing algorithms in ballistic laser gravimeters under the effect of correlated interference, the method of generalized least squares is applied. In this case, to describe the interference, a mathematical model of the autoregression process is used, for which the inverse correlation matrix has a band type and is expressed through the values of the autoregression coefficients. To convert the “path-time” data from the output of the coincidence circuit of ballistic laser gravimeters to a process uniform in time, their local quadratic interpolation is used. Algorithms for data processing in a ballistic gravimeter, developed on the basis of a method of weighted least squares using orthogonal Hahn polynomials, are considered. To implement a symmetric measurement method, the symmetric Hahn polynomials, characterized by one parameter, are used. The method of mathematical modelling is used to study the gain in the accuracy of measuring the gravitational acceleration by the synthesized algorithms in comparison with the algorithm based on the method of least squares. It is shown that auto seismic interference in ballistic laser gravimeters with a symmetric measurement method can be significantly reduced by using a mathematical model of the second-order autoregressive process in the method of generalized least squares. A comparative analysis of the characteristics of the algorithms developed using the method of generalized least squares, the method of weighted least squares and the method of ordinary least squares is carried out.


2020 ◽  
Vol 32 (22) ◽  
pp. 17077-17095 ◽  
Author(s):  
Stephanie Earp ◽  
Andrew Curtis

Abstract Travel-time tomography for the velocity structure of a medium is a highly nonlinear and nonunique inverse problem. Monte Carlo methods are becoming increasingly common choices to provide probabilistic solutions to tomographic problems but those methods are computationally expensive. Neural networks can often be used to solve highly nonlinear problems at a much lower computational cost when multiple inversions are needed from similar data types. We present the first method to perform fully nonlinear, rapid and probabilistic Bayesian inversion of travel-time data for 2D velocity maps using a mixture density network. We compare multiple methods to estimate probability density functions that represent the tomographic solution, using different sets of prior information and different training methodologies. We demonstrate the importance of prior information in such high-dimensional inverse problems due to the curse of dimensionality: unrealistically informative prior probability distributions may result in better estimates of the mean velocity structure; however, the uncertainties represented in the posterior probability density functions then contain less information than is obtained when using a less informative prior. This is illustrated by the emergence of uncertainty loops in posterior standard deviation maps when inverting travel-time data using a less informative prior, which are not observed when using networks trained on prior information that includes (unrealistic) a priori smoothness constraints in the velocity models. We show that after an expensive program of network training, repeated high-dimensional, probabilistic tomography is possible on timescales of the order of a second on a standard desktop computer.


2014 ◽  
Vol 505-506 ◽  
pp. 719-726 ◽  
Author(s):  
Tao Wen ◽  
Chang Cheng Li ◽  
Chun Jiang Che ◽  
Lian De Zhong ◽  
Xin Xin

Massive expressway toll data contained lots of valuable information. However, the skills of mining and analyzing toll data were limited currently. This study explored the modeling method of road network travel time reliability based on massive toll data. Firstly, this study obtained travel time data sample of each link at different months, and analyzed travel time statistical properties preliminarily. Secondly, this study used normal distribution, gamma distribution and Weibull distribution to fit travel time data sample, and different statistical indicators were involved to measure the fitting effect. Fitting results showed that normal distribution for link travel time was more rational and acceptable than the others. Thus, this study established link travel time reliability model, and proposed moment estimation method of calibrating the model parameters. In practical application, the reliability model can be used to judge traffic operating posture for expressway management department, and also can be used to forecast travel time information, to provide valuable reference on decision-making for drivers travel plan or route choice.


2011 ◽  
Author(s):  
A. J. Hunter ◽  
B. W. Drinkwater ◽  
P. D. Wilcox ◽  
Donald O. Thompson ◽  
Dale E. Chimenti

1998 ◽  
Vol 41 (1) ◽  
Author(s):  
M. ou A. Bounif ◽  
C. Dorbath

Local earthquake travel-time data were inverted to obtain a three dimensional tomographic image of the region centered on the 1985 Constantine earthquake. The resulting velocity model was then used to relocate the events. The tomographic data set consisted of P and S waves travel-times from 653 carefully selected aftershocks of this moderate size earthquake, recorded at 10 temporary stations. A three-dimensional P-wave velocity image to a depth of 12 km was obtained by Thurber's method. At shallower depth, the velocity contrasts reflected the differences in tectonic units. Velocities lower than 4 km/s corresponded to recent deposits, velocities higher than 5 km/s to the Constantine Neritic and the Tellian nappes. The relocation of the aftershocks indicates that most of the seismicity occured where the velocity exceeded 5.5 km/s. The aftershock distribution accurately defined the three segments involved in the main shock and led to a better understanding of the rupture process.


Sign in / Sign up

Export Citation Format

Share Document