scholarly journals Technical Note: The Normal Quantile Transformation and its application in a flood forecasting system

2011 ◽  
Vol 8 (5) ◽  
pp. 9275-9297 ◽  
Author(s):  
K. Bogner ◽  
F. Pappenberger ◽  
H. L. Cloke

Abstract. The Normal Quantile Transform (NQT) has been used in many hydrological and meteorological applications in order to make the Cumulated Density Function (CDF) of the observed, simulated and forecast river discharge, water level or precipitation data Gaussian. It is also the heart of the meta-Gaussian model for assessing the total predictive uncertainty of the Hydrological Uncertainty Processor (HUP) developed by Krzysztofowicz. In the field of geo-statistics this transformation is better known as Normal-Score Transform. In this paper some possible problems caused by small sample sizes for the applicability in flood forecasting systems will be discussed and illustrated by examples. For the practical implementation commands and examples from the freely available and widely used statistical computing language R (R Development Core Team, 2011) will be given (represented in Courier font) and possible solutions are suggested by combining extreme value analysis and non-parametric regression methods.

2012 ◽  
Vol 16 (4) ◽  
pp. 1085-1094 ◽  
Author(s):  
K. Bogner ◽  
F. Pappenberger ◽  
H. L. Cloke

Abstract. The Normal Quantile Transform (NQT) has been used in many hydrological and meteorological applications in order to make the Cumulated Distribution Function (CDF) of the observed, simulated and forecast river discharge, water level or precipitation data Gaussian. It is also the heart of the meta-Gaussian model for assessing the total predictive uncertainty of the Hydrological Uncertainty Processor (HUP) developed by Krzysztofowicz. In the field of geo-statistics this transformation is better known as the Normal-Score Transform. In this paper some possible problems caused by small sample sizes when applying the NQT in flood forecasting systems will be discussed and a novel way to solve the problem will be outlined by combining extreme value analysis and non-parametric regression methods. The method will be illustrated by examples of hydrological stream-flow forecasts.


Water ◽  
2018 ◽  
Vol 10 (11) ◽  
pp. 1519 ◽  
Author(s):  
Paul Muñoz ◽  
Johanna Orellana-Alvear ◽  
Patrick Willems ◽  
Rolando Célleri

Flash-flood forecasting has emerged worldwide due to the catastrophic socio-economic impacts this hazard might cause and the expected increase of its frequency in the future. In mountain catchments, precipitation-runoff forecasts are limited by the intrinsic complexity of the processes involved, particularly its high rainfall variability. While process-based models are hard to implement, there is a potential to use the random forest algorithm due to its simplicity, robustness and capacity to deal with complex data structures. Here a step-wise methodology is proposed to derive parsimonious models accounting for both hydrological functioning of the catchment (e.g., input data, representation of antecedent moisture conditions) and random forest procedures (e.g., sensitivity analyses, dimension reduction, optimal input composition). The methodology was applied to develop short-term prediction models of varying time duration (4, 8, 12, 18 and 24 h) for a catchment representative of the Ecuadorian Andes. Results show that the derived parsimonious models can reach validation efficiencies (Nash-Sutcliffe coefficient) from 0.761 (4-h) to 0.384 (24-h) for optimal inputs composed only by features accounting for 80% of the model’s outcome variance. Improvement in the prediction of extreme peak flows was demonstrated (extreme value analysis) by including precipitation information in contrast to the use of pure autoregressive models.


2020 ◽  
Vol 7 (1) ◽  
pp. 14-21
Author(s):  
Alexander von Glinski ◽  
Emre Yilmaz ◽  
Ryan Goodmanson ◽  
Clifford Pierre ◽  
Sven Frieler ◽  
...  

Abstract The purpose of this study was to identify the 30 most cited articles on hip arthroscopy and discuss their influence on recent surgical treatment. Due to advancements in hip arthroscopy, there is a widening spectrum of diagnostic and treatment indications. The purpose of this study was to identify the 30 most cited articles on hip arthroscopy and discuss their influence on contemporary surgical treatment. The Thomson Reuters Web of Science was used to identify the 30 most cited studies on hip arthroscopy between 1900 and 2018. These 30 articles generated 6152 citations with an average of 205.07 citations per item. Number of citations ranged from 146 to 461. Twenty-five out of the 30 papers were clinical cohort studies with a level of evidence between III and IV, encompassing 4348 patients. Four studies were reviewed (one including a technical note) and one a case report. We were able to identify the 30 most cited articles in the field of hip arthroscopy. Most articles were reported in high-impact journals, but reported small sample sizes in a retrospective setting. Prospective multi-arm cohort trials or randomized clinical trials represent opportunities for future studies.


2019 ◽  
Author(s):  
Lili Blumenberg ◽  
Emily Kawaler ◽  
MacIntosh Cornwell ◽  
Shaleigh Smith ◽  
Kelly Ruggles ◽  
...  

AbstractUnbiased assays such as shotgun proteomics and RNA-seq provide high-resolution molecular characterization of tumors. These assays measure molecules with highly varied distributions, making interpretation and hypothesis testing challenging. Samples with the most extreme measurements for a molecule can reveal the most interesting biological insights, yet are often excluded from analysis. Furthermore, rare disease subtypes are, by definition, underrepresented in cancer cohorts. To provide a strategy for identifying molecules aberrantly enriched in small sample cohorts, we present BlackSheep--a package for non-parametric description and differential analysis of genome-wide data, available at https://github.com/ruggleslab/blackSheep. BlackSheep is a complementary tool to other differential expression analysis methods that may be underpowered when analyzing small subgroups in a larger cohort.


2018 ◽  
Author(s):  
Colleen Molloy Farrelly

This study aims to confirm prior findings on the usefulness of topological data analysis (TDA) in the analysis of small samples, particularly focused on cohorts of profoundly gifted students, as well as explore the use of TDA-based regression methods for statistical modeling with small samples. A subset of the Gross sample is analyzed through supervised and unsupervised methods, including 16 and 17 individuals, respectively. Unsupervised learning confirmed prior results suggesting that evenly gifted and unevenly gifted subpopulations fundamentally differ. Supervised learning focused on predicting graduate school attendance and awards earned during undergraduate studies, and TDA-based logistic regression models were compared with more traditional machine learning models for logistic regression. Results suggest 1) that TDA-based methods are capable of handing small samples and seem more robust to the issues that arise in small samples than other machine learning methods and 2) that early childhood achievement scores and several factors related to childhood education interventions (such as early entry and radical acceleration) play a role in predicting key educational and professional achievements in adulthood. Possible new directions from this work include the use of TDA-based tools in the analysis of rare cohorts thus-far relegated to qualitative analytics or case studies, as well as potential exploration of early educational factors and adult-level achievement in larger populations of the profoundly gifted, particularly within the Study of Exceptional Talent and Talent Identification Program cohorts.


Author(s):  
Ryota Wada ◽  
Philip Jonathan ◽  
Takuji Waseda

Abstract Extreme value analysis of significant wave height using data from a single location often incurs large uncertainty due to small sample size. Including wave data from nearby locations increases sample size at the risk of introducing dependency between extreme events and hence violating modelling assumptions. In this work, we consider extreme value analysis of spatial wave data from the 109-year GOMOS wave hindcast for the Gulf of Mexico, seeking to incorporate the effects of spatial dependence in a simple but effective manner. We demonstrate that, for estimation of return values at a given location, incorporation of data from a circular disk region with radius of approximately 5° (long.-lat.), centred at the location of interest, provides an appropriate basis for extreme value analysis using the STM-E approach of Wada et al. (2018).


2019 ◽  
pp. 40-46 ◽  
Author(s):  
V.V. Savchenko ◽  
A.V. Savchenko

We consider the task of automated quality control of sound recordings containing voice samples of individuals. It is shown that in this task the most acute is the small sample size. In order to overcome this problem, we propose the novel method of acoustic measurements based on relative stability of the pitch frequency within a voice sample of short duration. An example of its practical implementation using aninter-periodic accumulation of a speech signal is considered. An experimental study with specially developed software provides statistical estimates of the effectiveness of the proposed method in noisy environments. It is shown that this method rejects the audio recording as unsuitable for a voice biometric identification with a probability of 0,95 or more for a signal to noise ratio below 15 dB. The obtained results are intended for use in the development of new and modifying existing systems of collecting and automated quality control of biometric personal data. The article is intended for a wide range of specialists in the field of acoustic measurements and digital processing of speech signals, as well as for practitioners who organize the work of authorized organizations in preparing for registration samples of biometric personal data.


Sign in / Sign up

Export Citation Format

Share Document