scholarly journals Parameter-dependent convergence bounds and complexity measure for a class of conceptual hydrological models

2011 ◽  
Vol 14 (2) ◽  
pp. 443-463 ◽  
Author(s):  
Saket Pande ◽  
Luis A. Bastidas ◽  
Sandjai Bhulai ◽  
Mac McKee

We provide analytical bounds on convergence rates for a class of hydrologic models and consequently derive a complexity measure based on the Vapnik–Chervonenkis (VC) generalization theory. The class of hydrologic models is a spatially explicit interconnected set of linear reservoirs with the aim of representing globally nonlinear hydrologic behavior by locally linear models. Here, by convergence rate, we mean convergence of the empirical risk to the expected risk. The derived measure of complexity measures a model's propensity to overfit data. We explore how data finiteness can affect model selection for this class of hydrologic model and provide theoretical results on how model performance on a finite sample converges to its expected performance as data size approaches infinity. These bounds can then be used for model selection, as the bounds provide a tradeoff between model complexity and model performance on finite data. The convergence bounds for the considered hydrologic models depend on the magnitude of their parameters, which are the recession parameters of constituting linear reservoirs. Further, the complexity of hydrologic models not only varies with the magnitude of their parameters but also depends on the network structure of the models (in terms of the spatial heterogeneity of parameters and the nature of hydrologic connectivity).

2017 ◽  
Vol 21 (2) ◽  
pp. 879-896 ◽  
Author(s):  
Tirthankar Roy ◽  
Hoshin V. Gupta ◽  
Aleix Serrat-Capdevila ◽  
Juan B. Valdes

Abstract. Daily, quasi-global (50° N–S and 180° W–E), satellite-based estimates of actual evapotranspiration at 0.25° spatial resolution have recently become available, generated by the Global Land Evaporation Amsterdam Model (GLEAM). We investigate the use of these data to improve the performance of a simple lumped catchment-scale hydrologic model driven by satellite-based precipitation estimates to generate streamflow simulations for a poorly gauged basin in Africa. In one approach, we use GLEAM to constrain the evapotranspiration estimates generated by the model, thereby modifying daily water balance and improving model performance. In an alternative approach, we instead change the structure of the model to improve its ability to simulate actual evapotranspiration (as estimated by GLEAM). Finally, we test whether the GLEAM product is able to further improve the performance of the structurally modified model. Results indicate that while both approaches can provide improved simulations of streamflow, the second approach also improves the simulation of actual evapotranspiration significantly, which substantiates the importance of making diagnostic structural improvements to hydrologic models whenever possible.


2021 ◽  
Author(s):  
Sophia Eugeni ◽  
Eric Vaags ◽  
Steven V. Weijs

<p>Accurate hydrologic modelling is critical to effective water resource management. As catchment attributes strongly influence the hydrologic behaviors in an area, they can be used to inform hydrologic models to better predict the discharge in a basin. Some basins may be more difficult to accurately predict than others. The difficulty in predicting discharge may also be related to the complexity of the discharge signal. The study establishes the relationship between a catchment’s static attributes and hydrologic model performance in those catchments, and also investigates the link to complexity, which we quantify with measures of compressibility based in information theory. </p><p>The project analyzes a large national dataset, comprised of catchment attributes for basins across the United States, paired with established performance metrics for corresponding hydrologic models. Principal Component Analysis (PCA) was completed on the catchment attributes data to determine the strongest modes in the input. The basins were clustered according to their catchment attributes and the performance within the clusters was compared. </p><p>Significant differences in model performance emerged between the clusters of basins. For the complexity analysis, details of the implementation and technical challenges will be discussed, as well as preliminary results.</p>


Water ◽  
2020 ◽  
Vol 12 (8) ◽  
pp. 2312
Author(s):  
Joseph A. Daraio

Hydrologic models driven by downscaled meteorologic data from general circulation models (GCM) should be evaluated using long-term simulations over a historical period. However, simulations driven by GCM data cannot be directly evaluated using observed flows, and the confidence in the results can be relatively low. The objectives of this paper were to bias correct simulated stream flows from calibrated hydrologic models for two basins in New Jersey, USA, and evaluate model performance in comparison to uncorrected simulations. Then, we used stream flow bias correction and flow duration curves (FDCs) to evaluate and assess simulations driven by statistically downscaled GCMs for the historical period and the future time slices 2041–2070 and 2071–2099. Bias correction of stream flow from simulations increased confidence in the performance of two previously calibrated hydrologic models. Results indicated there was no difference in projected FDCs for uncorrected and bias-corrected flows in one basin, while this was not the case in the second basin. This result provided greater confidence in projected stream flow changes in the former basin and implied more uncertainty in projected stream flows in the latter. Applications in water resources can use the methods described to evaluate the performance of GCM-driven simulations and assess the potential impacts of climate change with an appropriate level of confidence in the model results.


Author(s):  
J. Serrano ◽  
J. M. Jamilla ◽  
B. C. Hernandez ◽  
E. Herrera

Abstract. Runoffs from hydrologic models are often used in flood models, among other applications. These runoffs are converted from rainfall, signifying the importance of weather data accuracy. A common challenge for modelers is local weather data sparsity in most watersheds. Global weather datasets are often used as alternative. This study investigates the statistical significance and accuracy between using local weather data for hydrologic models and using the Climate Forecast System Reanalysis (CFSR), a global weather dataset. The Soil and Water Assessment Tool (SWAT) was used to compare the two weather data inputs in terms of generated discharges. Both long-term and event-based results were investigated to compare the models against absolute discharge values. The basin’s average total annual rainfall from the CFSR-based model (4062 mm) was around 1.5 times the local weather-based model (2683 mm). These basin precipitations yielded annual average flows of 53.4 cms and 26.7 cms for CFSR-based and local weather-based models, respectively. For the event-based scenario, the dates Typhoon Ketsana passed through the Philippine Area of Responsibility were checked. CFSR only read a spatially averaged maximum daily rainfall of 18.8 mm while the local gauges recorded 157.2 mm. Calibration and validation of the models were done using the observed discharges in Sto. Niño Station. The calibration of local weather-based model yielded satisfactory results for the Nash-Sutcliffe Efficiency (NSE), percent of bias (PBIAS), and ratio of the RMSE to the standard deviation of measured data (RSR). Meanwhile, the calibration of CFSR model yielded unsatisfactory values for all three parameters.


2013 ◽  
Vol 16 (3) ◽  
pp. 588-599 ◽  
Author(s):  
Kenneth J. Tobin ◽  
Marvin E. Bennett

With the proliferation of remote sensing platforms as well as numerous ground products based on weather radar estimation, there are now multiple options for precipitation data beyond traditional rain gauges for which most hydrologic models were originally designed. This study evaluates four precipitation products as input for generating streamflow simulations using two hydrologic models that significantly vary in complexity. The four precipitation products include two ground products from the National Weather Service: the Multi-sensor Precipitation Estimator (MPE) and rain gauge data. The two satellite products come from NASA's Tropical Rainfall Measurement Mission (TRMM) and include the TRMM 3B42 Research Version 6, which has a built-in ground bias correction, and the real-time TRMM Multi-Satellite Precipitation Analysis. The two hydrologic models utilized include the Soil and Water Assessment Tool (SWAT) and Gridded Surface and Subsurface Hydrologic Analysis (GSSHA). Simulations were conducted in three, moderate- to large-sized basins across the southern United States, the San Casimiro (South Texas), Skuna (northern Mississippi), Alapaha (southern Georgia), and were run for over 2 years. This study affirms the realization that input precipitation is at least as important as the choice of hydrologic model.


2020 ◽  
Author(s):  
Manuela I. Brunner ◽  
Lieke A. Melsen ◽  
Andrew W. Wood ◽  
Oldrich Rakovec ◽  
Naoki Mizukami ◽  
...  

Abstract. Floods cause large damages, especially if they affect large regions. Assessments of current, local and regional flood hazards and their future changes often involve the use of hydrologic models. However, uncertainties in simulated floods can be considerable and yield unreliable hazard and climate change impact assessments. A reliable hydrologic model ideally reproduces both local flood characteristics and spatial aspects of flooding, which is, however, not guaranteed especially when using standard model calibration metrics. In this paper we investigate how flood timing, magnitude and spatial variability are represented by an ensemble of hydrological models when calibrated on streamflow using the Kling–Gupta efficiency metric, an increasingly common metric of hydrologic model performance. We compare how four well-known models (SAC, HBV, VIC, and mHM) represent (1) flood characteristics and their spatial patterns; and (2) how they translate changes in meteorologic variables that trigger floods into changes in flood magnitudes. Our results show that both the modeling of local and spatial flood characteristics is challenging. They further show that changes in precipitation and temperature are not necessarily well translated to changes in flood flow, which makes local and regional flood hazard assessments even more difficult for future conditions. We conclude that models calibrated on integrated metrics such as the Kling–Gupta efficiency alone have limited reliability in flood hazard assessments, in particular in regional and future assessments, and suggest the development of alternative process-based and spatial evaluation metrics.


2016 ◽  
Vol 20 (11) ◽  
pp. 4655-4671 ◽  
Author(s):  
Steven L. Markstrom ◽  
Lauren E. Hay ◽  
Martyn P. Clark

Abstract. parameter hydrologic model, has been applied to the conterminous US (CONUS). Parameter sensitivity analysis was used to identify: (1) the sensitive input parameters and (2) particular model output variables that could be associated with the dominant hydrologic process(es). Sensitivity values of 35 PRMS calibration parameters were computed using the Fourier amplitude sensitivity test procedure on 110 000 independent hydrologically based spatial modeling units covering the CONUS and then summarized to process (snowmelt, surface runoff, infiltration, soil moisture, evapotranspiration, interflow, baseflow, and runoff) and model performance statistic (mean, coefficient of variation, and autoregressive lag 1). Identified parameters and processes provide insight into model performance at the location of each unit and allow the modeler to identify the most dominant process on the basis of which processes are associated with the most sensitive parameters. The results of this study indicate that: (1) the choice of performance statistic and output variables has a strong influence on parameter sensitivity, (2) the apparent model complexity to the modeler can be reduced by focusing on those processes that are associated with sensitive parameters and disregarding those that are not, (3) different processes require different numbers of parameters for simulation, and (4) some sensitive parameters influence only one hydrologic process, while others may influence many.


2020 ◽  
Author(s):  
Charles Luce ◽  
Abigail Lute

<p>A central question in model structural uncertainty is how complex a model should be in order to have greatest generality or transferability.  One school of thought is that models become more general by adding process subroutines.  On the other hand, model parameters and structures have been shown to change significantly when calibrated to different basins or time periods, suggesting that model complexity and model transferability may be antithetical.  An important facet to this discussion is noting that validation methods and data applied to model evaluation and selection may tend to bias answers to this question.  Here we apply non-random block cross-validation as a direct assessment of model transferability to a series of algorithmic space-time models of April 1 snow water equivalent (SWE) across 497 SNOTEL stations for 20 years.  In general, we show that low to moderate complexity models transfer most successfully to new conditions in space and time.  In other words, there is an optimum between overly complex and overly simple models.  Because structures in data resulting from temporal dynamics and spatial dependency in atmospheric and hydrological processes exist, naïvely applied cross-validation practices can lead to overfitting, overconfidence in model precision or reliability, and poor ability to infer causal mechanisms.  For example, random k-fold cross-validation methods, which are in common use for evaluating models, essentially assume independence of the data and would promote selection of more complex models.  We further demonstrate that blocks sampled with pseudoreplicated data can produce similar outcomes.  Some sampling strategies favored for hydrologic model validation may tend to promote pseudoreplication, requiring heightened attentiveness for model selection and evaluation.  While the illustrative examples are drawn from snow modeling, the concepts can be readily applied to common hydrologic modeling issues.</p>


2003 ◽  
Vol 15 (7) ◽  
pp. 1691-1714 ◽  
Author(s):  
Vladimir Cherkassky ◽  
Yunqian Ma

We discuss empirical comparison of analytical methods for model selection. Currently, there is no consensus on the best method for finite-sample estimation problems, even for the simple case of linear estimators. This article presents empirical comparisons between classical statistical methods—Akaike information criterion (AIC) and Bayesian information criterion (BIC)—and the structural risk minimization (SRM) method, basedon Vapnik-Chervonenkis (VC) theory, for regression problems. Our study is motivated by empirical comparisons in Hastie, Tibshirani, and Friedman (2001), which claims that the SRM method performs poorly for model selection and suggests that AIC yields superior predictive performance. Hence, we present empirical comparisons for various data sets and different types of estimators (linear, subset selection, and k-nearest neighbor regression). Our results demonstrate the practical advantages of VC-based model selection; it consistently outperforms AIC for all data sets. In our study, SRM and BIC methods show similar predictive performance. This discrepancy (between empirical results obtained using the same data) is caused by methodological drawbacks in Hastie et al. (2001), especially in their loose interpretation and application of SRM method. Hence, we discuss methodological issues important for meaningful comparisons and practical application of SRM method. We also point out the importance of accurate estimation of model complexity (VC-dimension) for empirical comparisons and propose a new practical estimate of model complexity for k-nearest neighbors regression.


Author(s):  
Adam Schreiner-McGraw ◽  
Hoori Ajami ◽  
Ray Anderson ◽  
Dong Wang

Accurate simulation of plant water use across agricultural ecosystems is essential for various applications, including precision agriculture, quantifying groundwater recharge, and optimizing irrigation rates. Previous approaches to integrating plant water use data into hydrologic models have relied on evapotranspiration (ET) observations. Recently, the flux variance similarity approach has been developed to partition ET to transpiration (T) and evaporation, providing an opportunity to use T data to parameterize models. To explore the value of T/ET data in improving hydrologic model performance, we examined multiple approaches to incorporate these observations for vegetation parameterization. We used ET observations from 5 eddy covariance towers located in the San Joaquin Valley, California, to parameterize orchard crops in an integrated land surface – groundwater model. We find that a simple approach of selecting the best parameter sets based on ET and T performance metrics works best at these study sites. Selecting parameters based on performance relative to observed ET creates an uncertainty of 27% relative to the observed value. When parameters are selected using both T and ET data, this uncertainty drops to 24%. Similarly, the uncertainty in potential groundwater recharge drops from 63% to 58% when parameters are selected with ET or T and ET data, respectively. Additionally, using crop type parameters results in similar levels of simulated ET as using site-specific parameters. Different irrigation schemes create high amounts of uncertainty and highlight the need for accurate estimates of irrigation when performing water budget studies.


Sign in / Sign up

Export Citation Format

Share Document