scholarly journals On accuracy of upper quantiles estimation

2010 ◽  
Vol 7 (4) ◽  
pp. 4761-4784
Author(s):  
I. Markiewicz ◽  
W. G. Strupczewski ◽  
K. Kochanek

Abstract. Flood frequency analysis (FFA) entails estimation of the upper tail of a probability density function (PDF) of annual peak flows obtained from either the annual maximum series or partial duration series. In hydrological practice the properties of various estimation methods of upper quantiles are identified with the case of known population distribution function. In reality the assumed hypothetical model differs from the true one and one can not assess the magnitude of error caused by model misspecification in respect to any estimated statistics. The opinion about the accuracy of the methods of upper quantiles estimation formed from the case of known population distribution function is upheld. The above-mentioned issue is the subject of the paper. The accuracy of large quantile assessments obtained from the four estimation methods are compared for two-parameter log-normal and log-Gumbel distributions and their three-parameter counterparts, i.e., three-parameter log-normal and GEV distributions. The cases of true and false hypothetical model are considered. The accuracy of flood quantile estimates depend on the sample size, on the distribution type, both true and hypothetical, and strongly depend on the estimation method. In particular, the maximum likelihood method looses its advantageous properties in case of model misspecification.

2010 ◽  
Vol 14 (11) ◽  
pp. 2167-2175 ◽  
Author(s):  
I. Markiewicz ◽  
W. G. Strupczewski ◽  
K. Kochanek

Abstract. Flood frequency analysis (FFA) entails the estimation of the upper tail of a probability density function (PDF) of annual peak flows obtained from either the annual maximum series or partial duration series. In hydrological practice, the properties of various methods of upper quantiles estimation are identified with the case of known population distribution function. In reality, the assumed hypothetical model differs from the true one and one cannot assess the magnitude of error caused by model misspecification in respect to any estimated statistics. The opinion about the accuracy of the methods of upper quantiles estimation formed from the case of known population distribution function is upheld. The above-mentioned issue is the subject of the paper. The accuracy of large quantile assessments obtained from the four estimation methods is compared to two-parameter log-normal and log-Gumbel distributions and their three-parameter counterparts, i.e., three-parameter log-normal and GEV distributions. The cases of true and false hypothetical models are considered. The accuracy of flood quantile estimates depends on the sample size, the distribution type (both true and hypothetical), and strongly depends on the estimation method. In particular, the maximum likelihood method loses its advantageous properties in case of model misspecification.


2021 ◽  
Author(s):  
Anne Bartens ◽  
Uwe Haberlandt

Abstract. In many cases flood frequency analysis needs to be carried out on mean daily flow (MDF) series without any available information on the instantaneous peak flow (IPF). We analyze the error of using MDFs instead of IPFs for flood quantile estimation on a German dataset and assess spatial patterns and factors that influence the deviation of MDF floods from their IPF counterparts. The main dependence could be found for catchment area but also gauge elevation appeared to have some influence. Based on the findings we propose simple linear models to correct both MDF flood peaks of individual flood events and overall MDF flood statistics. Key predictor in the models is the event-based ratio of flood peak and flood volume obtained directly from the daily flow records. This correction approach requires a minimum of data input, is easily applied, valid for the entire study area and successfully estimates IPF peaks and flood statistics. The models perform particularly well in smaller catchments, where other IPF estimation methods fall short. Still, the limit of the approach is reached for catchment sizes below 100 km2, where the hydrograph information from the daily series is no longer capable of approximating instantaneous flood dynamics.


2020 ◽  
Vol 45 (4) ◽  
pp. 475-506 ◽  
Author(s):  
Soojin Park ◽  
Gregory J. Palardy

Estimating the effects of randomized experiments and, by extension, their mediating mechanisms, is often complicated by treatment noncompliance. Two estimation methods for causal mediation in the presence of noncompliance have recently been proposed, the instrumental variable method (IV-mediate) and maximum likelihood method (ML-mediate). However, little research has examined their performance when certain assumptions are violated and under varying data conditions. This article addresses that gap in the research and compares the performance of the two methods. The results show that the distributional assumption of the compliance behavior plays an important role in estimation. That is, regardless of the estimation method or whether the other assumptions hold, results are biased if the distributional assumption is not met. We also found that the IV-mediate method is more sensitive to exclusion restriction violations, while the ML-mediate method is more sensitive to monotonicity violations. Moreover, estimates depend in part on compliance rate, sample size, and the availability and impact of control covariates. These findings are used to provide guidance on estimator selection.


2019 ◽  
Vol 1 (12) ◽  
Author(s):  
Mahmood Ul Hassan ◽  
Omar Hayat ◽  
Zahra Noreen

AbstractAt-site flood frequency analysis is a direct method of estimation of flood frequency at a particular site. The appropriate selection of probability distribution and a parameter estimation method are important for at-site flood frequency analysis. Generalized extreme value, three-parameter log-normal, generalized logistic, Pearson type-III and Gumbel distributions have been considered to describe the annual maximum steam flow at five gauging sites of Torne River in Sweden. To estimate the parameters of distributions, maximum likelihood estimation and L-moments methods are used. The performance of these distributions is assessed based on goodness-of-fit tests and accuracy measures. At most sites, the best-fitted distributions are with LM estimation method. Finally, the most suitable distribution at each site is used to predict the maximum flood magnitude for different return periods.


Water ◽  
2019 ◽  
Vol 11 (2) ◽  
pp. 193 ◽  
Author(s):  
Igor Leščešen ◽  
Dragan Dolinaj

In this paper, we performed Regional Flood Frequency Analysis (RFFA) by using L-moments and Annual Maximum Series (AMS) methods. Time series of volumes and duration of floods were derived using the threshold level method for 22 hydrological stations in the Pannonian Basin. For flood definition, a threshold set at Q10 was used. The aim of this research is to derive best-fit regional distribution for the four major rivers within the Pannonian Basin and to provide reliable prediction of flood quantiles. The results show that the investigated area can be considered homogeneous (Vi < 1) both for flood volumes (0.097) and durations (0.074). To determine the best-fit regional distribution, the six most commonly used distributions were used. Results obtained by L-moment ratio diagram and Z statistics show that all distributions satisfy the test criteria, but because the Log-Normal distribution has the value closest to zero, it can be selected as the best-fit distribution for the volumes (0.12) and durations (0.25) of floods.


2017 ◽  
Vol 42 (5) ◽  
pp. 359-375 ◽  
Author(s):  
Leah M. Feuerstahler

In item response theory (IRT), item response probabilities are a function of item characteristics and latent trait scores. Within an IRT framework, trait score misestimation results from (a) random error, (b) the trait score estimation method, (c) errors in item parameter estimation, and (d) model misspecification. This study investigated the relative effects of these error sources on the bias and confidence interval coverage rates for trait scores. Our results showed that overall, bias values were close to 0, and coverage rates were fairly accurate for central trait scores and trait estimation methods that did not use a strong Bayesian prior. However, certain types of model misspecifications were found to produce severely biased trait estimates with poor coverage rates, especially at extremes of the latent trait continuum. It is demonstrated that biased trait estimates result from estimated item response functions (IRFs) that exhibit systematic conditional bias, and that these conditionally biased IRFs may not be detected by model or item fit indices. One consequence of these results is that certain types of model misspecifications can lead to estimated trait scores that are nonlinearly related to the data-generating latent trait. Implications for item and trait score estimation and interpretation are discussed.


2019 ◽  
Vol 11 (4) ◽  
pp. 966-979
Author(s):  
Nur Amalina Mat Jan ◽  
Ani Shabri ◽  
Ruhaidah Samsudin

Abstract Non-stationary flood frequency analysis (NFFA) plays an important role in addressing the issue of the stationary assumption (independent and identically distributed flood series) that is no longer valid in infrastructure-designed methods. This confirms the necessity of developing new statistical models in order to identify the change of probability functions over time and obtain a consistent flood estimation method in NFFA. The method of Trimmed L-moments (TL-moments) with time covariate is confronted with the L-moment method for the stationary and non-stationary generalized extreme value (GEV) models. The aims of the study are to investigate the behavior of the proposed TL-moments method in the presence of NFFA and applying the method along with GEV distribution. Comparisons of the methods are made by Monte Carlo simulations and bootstrap-based method. The simulation study showed the better performance of most levels of TL-moments method, which is TL(η,0), (η = 2, 3, 4) than the L-moment method for all models (GEV1, GEV2, and GEV3). The TL-moment method provides more efficient quantile estimates than other methods in flood quantiles estimated at higher return periods. Thus, the TL-moments method can produce better estimation results since the L-moment eliminates lowest value and gives more weight to the largest value which provides important information.


Sign in / Sign up

Export Citation Format

Share Document