HydroNets: Leveraging River Network Structure and Deep Neural Networks for Hydrologic Modeling

Author(s):  
Zach Moshe ◽  
Asher Metzger ◽  
Frederik Kratzert ◽  
Efrat Morin ◽  
Sella Nevo ◽  
...  

<p>Accurate and scalable hydrologic models are essential building blocks of several important applications, from water resource management to timely flood warnings. In this work we present a novel family of hydrologic models, called HydroNets, that leverages river network connectivity structure within deep neural architectures. The injection of this connectivity structure prior knowledge allows for scalable and accurate hydrologic modeling.</p><p>Prior knowledge plays an important role in machine learning and AI. On one extreme of the prior knowledge spectrum there are expert systems, which exclusively rely on domain expertise encoded into a model. On the other extreme there are general purpose agnostic machine learning methods, which are exclusively data-driven, without intentional utilization of inductive bias for the problem at hand. In the context of hydrologic modeling, conceptual models such as the Sacramento Soil Moisture Accounting Model (SAC-SMA) are closer to expert systems. Such models require explicit functional modeling of water volume flow in terms of their input variables and model parameters (e.g., precipitation, hydraulic conductivity, etc.) which could be calibrated using data. Instances of agnostic methods for stream flow hydrologic modelling, which for the most part do not utilize problem specific bias, have recently been presented by Kratzert et al. (2018, 2019) and by Shalev et al. (2019). These works showed that general purpose deep recurrent neural networks, such as long short-term models (LSTMs), can achieve state-of-the-art hydrologic forecasts at scale with less information.</p><p>One of the fundamental reasons for the success of deep neural architectures in most application domains is the incorporation of prior knowledge into the architecture itself. This is, for example, the case in machine vision where convolutional layers and max pooling manifest essential invariances of visual perception. In this work we present HydroNets, a family of neural network models for hydrologic forecasting. HydroNets leverage the inherent (graph-theoretic) tree structure of river water flow, existing in any multi-site hydrologic basin. The network architecture itself reflects river network connectivity and catchment structures such that each sub-basin is represented as a tree node, and edges represent water flow from sub-basins to their containing basin. HydroNets are constructed such that all nodes utilize a shared global model component, as well as site-specific sub-models for local modulations. HydroNets thus combine two signals: site specific rainfall-runoff and upstream network dynamics, which can lead to improved predictions at longer horizons. Moreover, the proposed architecture, with its shared global model, tend to reduce sample complexity, increase scalability, and allows for transferability to sub-basins that suffer from scarce historical data. We present several simulation results over multiple basins in both India and the USA that convincingly support the proposed model and its advantages.</p>

1995 ◽  
Vol 3 ◽  
pp. 147-185 ◽  
Author(s):  
C. G. Giraud-Carrier ◽  
T. R. Martinez

Learning and reasoning are both aspects of what is considered to be intelligence. Their studies within AI have been separated historically, learning being the topic of machine learning and neural networks, and reasoning falling under classical (or symbolic) AI. However, learning and reasoning are in many ways interdependent. This paper discusses the nature of some of these interdependencies and proposes a general framework called FLARE, that combines inductive learning using prior knowledge together with reasoning in a propositional setting. Several examples that test the framework are presented, including classical induction, many important reasoning protocols and two simple expert systems.


2020 ◽  
Vol 25 (2) ◽  
pp. 7-13
Author(s):  
Zhangozha A.R. ◽  

On the example of the online game Akinator, the basic principles on which programs of this type are built are considered. Effective technics have been proposed by which artificial intelligence systems can build logical inferences that allow to identify an unknown subject from its description (predicate). To confirm the considered hypotheses, the terminological analysis of definition of the program "Akinator" offered by the author is carried out. Starting from the assumptions given by the author's definition, the article complements their definitions presented by other researchers and analyzes their constituent theses. Finally, some proposals are made for the next steps in improving the program. The Akinator program, at one time, became one of the most famous online games using artificial intelligence. And although this was not directly stated, it was clear to the experts in the field of artificial intelligence that the program uses the techniques of expert systems and is built on inference rules. At the moment, expert systems have lost their positions in comparison with the direction of neural networks in the field of artificial intelligence, however, in the case considered in the article, we are talking about techniques using both directions – hybrid systems. Games for filling semantics interact with the user, expanding their semantic base (knowledge base) and use certain strategies to achieve the best result. The playful form of such semantics filling programs is beneficial for researchers by involving a large number of players. The article examines the techniques used by the Akinator program, and also suggests possible modifications to it in the future. This study, first of all, focuses on how the knowledge base of the Akinator program is built, it consists of incomplete sets, which can be filled and adjusted as a result of further iterations of the program launches. It is important to note our assumption that the order of questions used by the program during the game plays a key role, because it determines its strategy. It was identified that the program is guided by the principles of nonmonotonic logic – the assumptions constructed by the program are not final and can be rejected by it during the game. The three main approaches to acquisite semantics proposed by Jakub Šimko and Mária Bieliková are considered, namely, expert work, crowdsourcing and machine learning. Paying attention to machine learning, the Akinator program using machine learning to build an effective strategy in the game presents a class of hybrid systems that combine the principles of two main areas in artificial intelligence programs – expert systems and neural networks.


2012 ◽  
pp. 1404-1416 ◽  
Author(s):  
David Parry

Decision analysis techniques attempt to utilize mathematical data about outcomes and preferences to help people make optimal decisions. The increasing uses of computerized records and powerful computers have made these techniques much more accessible and usable. The partnership between women and clinicians can be enhanced by sharing information, knowledge, and the decision making process in this way. Other techniques for assisting with decision making, such as learning from data via neural networks or other machine learning approaches may offer increased value. Rules learned from such approaches may allow the development of expert systems that actually take over some of the decision making role, although such systems are not yet in widespread use.


2019 ◽  
Vol 28 (01) ◽  
pp. 027-034 ◽  
Author(s):  
Laszlo Balkanyi ◽  
Ronald Cornet

Introduction: Artificial intelligence (AI) is widespread in many areas, including medicine. However, it is unclear what exactly AI encompasses. This paper aims to provide an improved understanding of medical AI and its constituent fields, and their interplay with knowledge representation (KR). Methods: We followed a Wittgensteinian approach (“meaning by usage”) applied to content metadata labels, using the Medical Subject Headings (MeSH) thesaurus to classify the field. To understand and characterize medical AI and the role of KR, we analyzed: (1) the proportion of papers in MEDLINE related to KR and various AI fields; (2) the interplay among KR and AI fields and overlaps among the AI fields; (3) interconnectedness of fields; and (4) phrase frequency and collocation based on a corpus of abstracts. Results: Data from over eighty thousand papers showed a steep, six-fold surge in the last 30 years. This growth happened in an escalating and cascading way. A corpus of 246,308 total words containing 21,842 unique words showed several hundred occurrences of notions such as robotics, fuzzy logic, neural networks, machine learning and expert systems in the phrase frequency analysis. Collocation analysis shows that fuzzy logic seems to be the most often collocated notion. Neural networks and machine learning are also used in the conceptual neighborhood of KR. Robotics is more isolated. Conclusions: Authors note an escalation of published AI studies in medicine. Knowledge representation is one of the smaller areas, but also the most interconnected, and provides a common cognitive layer for other areas.


Author(s):  
David Parry

Decision analysis techniques attempt to utilize mathematical data about outcomes and preferences to help people make optimal decisions. The increasing uses of computerized records and powerful computers have made these techniques much more accessible and usable. The partnership between women and clinicians can be enhanced by sharing information, knowledge, and the decision making process in this way. Other techniques for assisting with decision making, such as learning from data via neural networks or other machine learning approaches may offer increased value. Rules learned from such approaches may allow the development of expert systems that actually take over some of the decision making role, although such systems are not yet in widespread use.


Information ◽  
2019 ◽  
Vol 10 (3) ◽  
pp. 98 ◽  
Author(s):  
Tariq Ahmad ◽  
Allan Ramsay ◽  
Hanady Ahmed

Assigning sentiment labels to documents is, at first sight, a standard multi-label classification task. Many approaches have been used for this task, but the current state-of-the-art solutions use deep neural networks (DNNs). As such, it seems likely that standard machine learning algorithms, such as these, will provide an effective approach. We describe an alternative approach, involving the use of probabilities to construct a weighted lexicon of sentiment terms, then modifying the lexicon and calculating optimal thresholds for each class. We show that this approach outperforms the use of DNNs and other standard algorithms. We believe that DNNs are not a universal panacea and that paying attention to the nature of the data that you are trying to learn from can be more important than trying out ever more powerful general purpose machine learning algorithms.


Author(s):  
M. Kada ◽  
D. Kuramin

Abstract. In the practical and professional work of classifying airborne laser scanning (ALS) point clouds, there are nowadays numerous methods and software applications available that are able to separate the points into a few basic categories and do so with a known and consistent quality. Further refinement of the classes then requires either manual or semi-automatic work, or the use of supervised machine learning algorithms. In using supervised machine learning, e.g. Deep Learning neural networks, however, there is a significant chance that they will not maintain the approved quality of an existing classification. In this study, we therefore evaluate the application of two neural networks, PointNet++ and KPConv, and propose to integrate prior knowledge from a pre-existing classification in the form of height above ground and an encoding of the already available labels as additional per-point input features. Our experiments show that such an approach can improve the quality of the 3D classification results by 6% to 10% in mean intersection over union (mIoU) depending on the respective network, but it also cannot completely avoid the aforementioned problems.


2018 ◽  
Vol 35 (2) ◽  
pp. e12258
Author(s):  
José García-Rodríguez ◽  
Sergio Escalera ◽  
Alexandra Psarrou ◽  
Isabelle Guyon ◽  
Andrew Lewis ◽  
...  

2020 ◽  
Vol 2020 ◽  
Author(s):  
Chaya Liebeskind ◽  
Shmuel Liebeskind

In this study, we address the interesting task of classifying historical texts by their assumed period of writ-ing. This task is useful in digital humanity studies where many texts have unidentified publication dates.For years, the typical approach for temporal text classification was supervised using machine-learningalgorithms. These algorithms require careful feature engineering and considerable domain expertise todesign a feature extractor to transform the raw text into a feature vector from which the classifier couldlearn to classify any unseen valid input. Recently, deep learning has produced extremely promising re-sults for various tasks in natural language processing (NLP). The primary advantage of deep learning isthat human engineers did not design the feature layers, but the features were extrapolated from data witha general-purpose learning procedure. We investigated deep learning models for period classification ofhistorical texts. We compared three common models: paragraph vectors, convolutional neural networks (CNN) and recurrent neural networks (RNN), and conventional machine-learning methods. We demon-strate that the CNN and RNN models outperformed the paragraph vector model and the conventionalsupervised machine-learning algorithms. In addition, we constructed word embeddings for each timeperiod and analyzed semantic changes of word meanings over time.


Sign in / Sign up

Export Citation Format

Share Document