scholarly journals Technical note: “Bit by bit”: a practical and general approach for evaluating model computational complexity vs. model performance

2021 ◽  
Vol 25 (2) ◽  
pp. 1103-1115
Author(s):  
Elnaz Azmi ◽  
Uwe Ehret ◽  
Steven V. Weijs ◽  
Benjamin L. Ruddell ◽  
Rui A. P. Perdigão

Abstract. One of the main objectives of the scientific enterprise is the development of well-performing yet parsimonious models for all natural phenomena and systems. In the 21st century, scientists usually represent their models, hypotheses, and experimental observations using digital computers. Measuring performance and parsimony of computer models is therefore a key theoretical and practical challenge for 21st century science. “Performance” here refers to a model's ability to reduce predictive uncertainty about an object of interest. “Parsimony” (or complexity) comprises two aspects: descriptive complexity – the size of the model itself which can be measured by the disk space it occupies – and computational complexity – the model's effort to provide output. Descriptive complexity is related to inference quality and generality; computational complexity is often a practical and economic concern for limited computing resources. In this context, this paper has two distinct but related goals. The first is to propose a practical method of measuring computational complexity by utility software “Strace”, which counts the total number of memory visits while running a model on a computer. The second goal is to propose the “bit by bit” method, which combines measuring computational complexity by “Strace” and measuring model performance by information loss relative to observations, both in bit. For demonstration, we apply the “bit by bit” method to watershed models representing a wide diversity of modelling strategies (artificial neural network, auto-regressive, process-based, and others). We demonstrate that computational complexity as measured by “Strace” is sensitive to all aspects of a model, such as the size of the model itself, the input data it reads, its numerical scheme, and time stepping. We further demonstrate that for each model, the bit counts for computational complexity exceed those for performance by several orders of magnitude and that the differences among the models for both computational complexity and performance can be explained by their setup and are in accordance with expectations. We conclude that measuring computational complexity by “Strace” is practical, and it is also general in the sense that it can be applied to any model that can be run on a digital computer. We further conclude that the “bit by bit” approach is general in the sense that it measures two key aspects of a model in the single unit of bit. We suggest that it can be enhanced by additionally measuring a model's descriptive complexity – also in bit.

2020 ◽  
Author(s):  
Elnaz Azmi ◽  
Uwe Ehret ◽  
Steven V. Weijs ◽  
Benjamin L. Ruddell ◽  
Rui A. P. Perdigão

Abstract. One of the main objectives of the scientific enterprise is the development of parsimonious yet well-performing models for all natural phenomena and systems. In the 21st century, scientists usually represent their models, hypotheses, and experimental observations using digital computers. Measuring performance and parsimony for computer models is therefore a key theoretical and practical challenge for 21st century science. The basic dimensions of computer model parsimony are descriptive complexity, i.e. the length of the model itself, and computational complexity, i.e. the model's effort to provide output. Descriptive complexity is related to inference quality and generality, and Occam's razor advocates minimizing this complexity. Computational complexity is a practical and economic concern for limited computing resources. Both complexities measure facets of the phenomenological or natural complexity of the process or system that is being observed, analysed and modelled. This paper presents a practical technique for measuring the computational complexity of a digital dynamical model and its performance bit by bit. Computational complexity is measured by the average number of memory visits per simulation time step in bits, and model performance is expressed by its inverse, information loss, measured by conditional entropy of observations given the related model predictions, also in bits. We demonstrate this technique by applying it to a variety of watershed models representing a wide diversity of modelling strategies including artificial neural network, auto-regressive, simple and more advanced process-based, and both approximate and exact restatements of experimental observations. Comparing the models revealed that the auto-regressive model poses a favourable trade-off with high performance and low computational complexity, but neural networks and high-time-frequency conceptual bucket models pose an unfavourable trade-off with low performance and high computational complexity. We conclude that the bit by bit approach is a practical approach for evaluating models in terms of performance and computational complexity, both in the universal unit of bits, which also can be used to express the other main aspect of model parsimony, description length.


2019 ◽  
Vol 23 (10) ◽  
pp. 4323-4331 ◽  
Author(s):  
Wouter J. M. Knoben ◽  
Jim E. Freer ◽  
Ross A. Woods

Abstract. A traditional metric used in hydrology to summarize model performance is the Nash–Sutcliffe efficiency (NSE). Increasingly an alternative metric, the Kling–Gupta efficiency (KGE), is used instead. When NSE is used, NSE = 0 corresponds to using the mean flow as a benchmark predictor. The same reasoning is applied in various studies that use KGE as a metric: negative KGE values are viewed as bad model performance, and only positive values are seen as good model performance. Here we show that using the mean flow as a predictor does not result in KGE = 0, but instead KGE =1-√2≈-0.41. Thus, KGE values greater than −0.41 indicate that a model improves upon the mean flow benchmark – even if the model's KGE value is negative. NSE and KGE values cannot be directly compared, because their relationship is non-unique and depends in part on the coefficient of variation of the observed time series. Therefore, modellers who use the KGE metric should not let their understanding of NSE values guide them in interpreting KGE values and instead develop new understanding based on the constitutive parts of the KGE metric and the explicit use of benchmark values to compare KGE scores against. More generally, a strong case can be made for moving away from ad hoc use of aggregated efficiency metrics and towards a framework based on purpose-dependent evaluation metrics and benchmarks that allows for more robust model adequacy assessment.


Water ◽  
2021 ◽  
Vol 13 (15) ◽  
pp. 2032
Author(s):  
Pâmela A. Melo ◽  
Lívia A. Alvarenga ◽  
Javier Tomasella ◽  
Carlos R. Mello ◽  
Minella A. Martins ◽  
...  

Landform classification is important for representing soil physical properties varying continuously across the landscape and for understanding many hydrological processes in watersheds. Considering it, this study aims to use a geomorphology map (Geomorphons) as an input to a physically based hydrological model (Distributed Hydrology Soil Vegetation Model (DHSVM)) in a mountainous headwater watershed. A sensitivity analysis of five soil parameters was evaluated for streamflow simulation in each Geomorphons feature. As infiltration and saturation excess overland flow are important mechanisms for streamflow generation in complex terrain watersheds, the model’s input soil parameters were most sensitive in the “slope”, “hollow”, and “valley” features. Thus, the simulated streamflow was compared with observed data for calibration and validation. The model performance was satisfactory and equivalent to previous simulations in the same watershed using pedological survey and moisture zone maps. Therefore, the results from this study indicate that a geomorphologically based map is applicable and representative for spatially distributing hydrological parameters in the DHSVM.


Data ◽  
2021 ◽  
Vol 6 (6) ◽  
pp. 60
Author(s):  
Miguel A. Becerra ◽  
Catalina Tobón ◽  
Andrés Eduardo Castro-Ospina ◽  
Diego H. Peluffo-Ordóñez

This paper provides a comprehensive description of the current literature on data fusion, with an emphasis on Information Quality (IQ) and performance evaluation. This literature review highlights recent studies that reveal existing gaps, the need to find a synergy between data fusion and IQ, several research issues, and the challenges and pitfalls in this field. First, the main models, frameworks, architectures, algorithms, solutions, problems, and requirements are analyzed. Second, a general data fusion engineering process is presented to show how complex it is to design a framework for a specific application. Third, an IQ approach, as well as the different methodologies and frameworks used to assess IQ in information systems are addressed; in addition, data fusion systems are presented along with their related criteria. Furthermore, information on the context in data fusion systems and its IQ assessment are discussed. Subsequently, the issue of data fusion systems’ performance is reviewed. Finally, some key aspects and concluding remarks are outlined, and some future lines of work are gathered.


Water ◽  
2021 ◽  
Vol 13 (11) ◽  
pp. 1456
Author(s):  
Kee-Won Seong ◽  
Jang Hyun Sung

An oscillatory S-curve causes unexpected fluctuations in a unit hydrograph (UH) of desired duration or an instantaneous UH (IUH) that may affect the constraints for hydrologic stability. On the other hand, the Savitzky–Golay smoothing and differentiation filter (SG filter) is a digital filter known to smooth data without distorting the signal tendency. The present study proposes a method based on the SG filter to cope with oscillatory S-curves. Compared to previous conventional methods, the application of the SG filter to an S-curve was shown to drastically reduce the oscillation problems on the UH and IUH. In this method, the SG filter parameters are selected to give the minimum influence on smoothing and differentiation. Based on runoff reproduction results and performance criteria, it appears that the SG filter performed both smoothing and differentiation without the remarkable variation of hydrograph properties such as peak or time-to peak. The IUH, UH, and S-curve were estimated using storm data from two watersheds. The reproduced runoffs showed high levels of model performance criteria. In addition, the analyses of two other watersheds revealed that small watershed areas may experience scale problems. The proposed method is believed to be valuable when error-prone data are involved in analyzing the linear rainfall–runoff relationship.


2017 ◽  
Vol 17 (2) ◽  
pp. 185-196
Author(s):  
Mario Scalas ◽  
Palmalisa Marra ◽  
Luca Tedesco ◽  
Raffaele Quarta ◽  
Emanuele Cantoro ◽  
...  

Abstract. This article describes the architecture of sea situational awareness (SSA) platform, a major asset within TESSA, an industrial research project funded by the Italian Ministry of Education and Research. The main aim of the platform is to collect, transform and provide forecast and observational data as information suitable for delivery across a variety of channels, like web and mobile; specifically, the ability to produce and provide forecast information suitable for creating SSA-enabled applications has been a critical driving factor when designing and evolving the whole architecture. Thus, starting from functional and performance requirements, the platform architecture is described in terms of its main building blocks and flows among them: front-end components that support end-user applications and map and data analysis components that allow for serving maps and querying data. Focus is directed to key aspects and decisions about the main issues faced, like interoperability, scalability, efficiency and adaptability, but it also considers insights about future works in this and similarly related subjects. Some analysis results are also provided in order to better characterize critical issues and related solutions.


2021 ◽  
Author(s):  
Brian C. McFall ◽  
Douglas R. Krafft ◽  
Hande McCaw ◽  
Brooke M. Walker

This Regional Sediment Management Technical Note (RSM TN) provides practical metrics of success for nearshore nourishment projects constructed with dredged sediment. Clearly defined goals and performance metrics for projects will set clear expectations and will lead to longterm project support from local stakeholders and the public.


2020 ◽  
pp. 1-10
Author(s):  
Ankit Rawat ◽  
Mohd Fazle Azeem

The modeling of BLDC motor and performance analysis under diverse operating speed settings has been presented in this paper. BLDC motors gaining more & more attention from different Industrial and domestic appliance manufacturers due to its compact size, high efficiency and robust structure. Voluminous research and developments in the domains of material science and power electronics led to substantial increase in applications of BLDC motor to electric drives. This paper deals with the modeling of BLDC motor drive system along with a comparative study of modified queens bee evolution based GA tuned & manually tuned control schemes using MATLAB /SIMULINK. In order to evaluate the performance of proposed drive, simulation is carried out at different Mechanical load & speed conditions. Test outcomes thus achieved show that the model performance is satisfactory.


Sign in / Sign up

Export Citation Format

Share Document