Modeling of groundwater table depth anomalies using Long Short-Term Memory networks over Europe

Author(s):  
Yueling Ma ◽  
Carsten Montzka ◽  
Bagher Bayat ◽  
Stefan Kollet

<p>Groundwater is the dominant source of fresh water in many European countries. However, due to a lack of near-real-time water table depth (wtd) observations, monitoring of groundwater resources is not feasible at the continental scale. Thus, an alternative approach is required to produce wtd data from other available observations near-real-time. In this study, we propose Long Short-Term Memory (LSTM) networks to model monthly wtd anomalies over Europe utilizing monthly precipitation anomalies as input. LSTM networks are a special type of artificial neural networks, showing great promise in exploiting long-term dependencies between time series, which is expected in the response of groundwater to precipitation. To establish the methodology, spatially and temporally continuous data from terrestrial simulations at the continental scale were applied with a spatial resolution of 0.11°, ranging from the year 1996 to 2016 (Furusho-Percot et al., 2019). They were divided into a training set (1996 – 2012), a validation set (2012 – 2014) and a testing set (2015 -2016) to construct local models on selected pixels over eight PRUDENCE regions. The outputs of the LSTM networks showed good agreement with the simulation results in locations with a shallow wtd (~3m). It is important to note, the quality of the models was strongly affected by the amount of snow cover. Moreover, with the introduction of monthly evapotranspiration anomalies as additional input, pronounced improvements of the network performances were only obtained in more arid regions (i.e., Iberian Peninsula and Mediterranean). Our results demonstrate the potential of LSTM networks to produce high-quality wtd anomalies from hydrometeorological variables that are monitored at the large scale and part of operational forecasting systems potentially facilitating the implementation of an efficient groundwater monitoring system over Europe.</p><p>Reference:</p><p>Furusho-Percot, C., Goergen, K., Hartick, C., Kulkarni, K., Keune, J. and Kollet, S. (2019). Pan-European groundwater to atmosphere terrestrial systems climatology from a physically consistent simulation. Scientific Data, 6(1).</p>

2021 ◽  
Author(s):  
Yueling Ma ◽  
Carsten Montzka ◽  
Bagher Bayat ◽  
Stefan Kollet

<p>Near real-time groundwater table depth measurements are scarce over Europe, leading to challenges in monitoring groundwater resources at the continental scale. In this study, we leveraged knowledge learned from simulation results by Long Short-Term Memory (LSTM) networks to estimate monthly groundwater table depth anomaly (<em>wtd<sub>a</sub></em>) data over Europe. The LSTM networks were trained, validated, and tested at individual pixels on anomaly data derived from daily integrated hydrologic simulation results over Europe from 1996 to 2016, with a spatial resolution of 0.11° (Furusho-Percot et al., 2019), to predict monthly <em>wtd<sub>a</sub></em> based on monthly precipitation anomalies (<em>pr<sub>a</sub></em>) and soil moisture anomalies (<em>θ<sub>a</sub></em>). Without additional training, we directly fed the networks with averaged monthly <em>pr<sub>a</sub></em> and <em>θ<sub>a</sub></em> data from 1996 to 2016 obtained from commonly available observational datasets and reanalysis products, and compared the network outputs with available borehole <em>in situ</em> measured <em>wtd<sub>a</sub></em>. The LSTM network estimates show good agreement with the <em>in situ</em> observations, resulting in Pearson correlation coefficients of regional averaged <em>wtd<sub>a</sub></em> data in seven PRUDENCE regions ranging from 42% to 76%, which are ~ 10% higher than the original simulation results except for the Iberian Peninsula. Our study demonstrates the potential of LSTM networks to transfer knowledge from simulation to reality for the estimation of <em>wtd<sub>a</sub></em> over Europe. The proposed method can be used to provide spatiotemporally continuous information at large spatial scales in case of sparse ground-based observations, which is common for groundwater table depth measurements. Moreover, the results highlight the advantage of combining physically-based models with machine learning techniques in data processing.</p><p> </p><p>Reference:</p><p>Furusho-Percot, C., Goergen, K., Hartick, C., Kulkarni, K., Keune, J. and Kollet, S. (2019). Pan-European groundwater to atmosphere terrestrial systems climatology from a physically consistent simulation. Scientific Data, 6(1).</p>


2021 ◽  
Vol 3 ◽  
Author(s):  
Yueling Ma ◽  
Carsten Montzka ◽  
Bagher Bayat ◽  
Stefan Kollet

The lack of high-quality continental-scale groundwater table depth observations necessitates developing an indirect method to produce reliable estimation for water table depth anomalies (wtda) over Europe to facilitate European groundwater management under drought conditions. Long Short-Term Memory (LSTM) networks are a deep learning technology to exploit long-short-term dependencies in the input-output relationship, which have been observed in the response of groundwater dynamics to atmospheric and land surface processes. Here, we introduced different input variables including precipitation anomalies (pra), which is the most common proxy of wtda, for the networks to arrive at improved wtda estimates at individual pixels over Europe in various experiments. All input and target data involved in this study were obtained from the simulated TSMP-G2A data set. We performed wavelet coherence analysis to gain a comprehensive understanding of the contributions of different input variable combinations to wtda estimates. Based on the different experiments, we derived an indirect method utilizing LSTM networks with pra and soil moisture anomaly (θa) as input, which achieved the optimal network performance. The regional medians of test R2 scores and RMSEs obtained by the method in the areas with wtd ≤ 3.0 m were 76–95% and 0.17–0.30, respectively, constituting a 20–66% increase in median R2 and a 0.19–0.30 decrease in median RMSEs compared to the LSTM networks only with pra as input. Our results show that introducing θa significantly improved the performance of the trained networks to predict wtda, indicating the substantial contribution of θa to explain groundwater anomalies. Also, the European wtda map reproduced by the method had good agreement with that derived from the TSMP-G2A data set with respect to drought severity, successfully detecting ~41% of strong drought events (wtda ≥ 1.5) and ~29% of extreme drought events (wtda ≥ 2) in August 2015. The study emphasizes the importance to combine soil moisture information with precipitation information in quantifying or predicting groundwater anomalies. In the future, the indirect method derived in this study can be transferred to real-time monitoring of groundwater drought at the continental scale using remotely sensed soil moisture and precipitation observations or respective information from weather prediction models.


Hydrology ◽  
2020 ◽  
Vol 7 (3) ◽  
pp. 64 ◽  
Author(s):  
Mun-Ju Shin ◽  
Soo-Hyoung Moon ◽  
Kyung Goo Kang ◽  
Duk-Chul Moon ◽  
Hyuk-Joon Koh

To properly manage the groundwater resources, it is necessary to analyze the impact of groundwater withdrawal on the groundwater level. In this study, a Long Short-Term Memory (LSTM) network was used to evaluate the groundwater level prediction performance and analyze the impact of the change in the amount of groundwater withdrawal from the pumping wells on the change in the groundwater level in the nearby monitoring wells located in Jeju Island, Korea. The Nash–Sutcliffe efficiency between the observed and simulated groundwater level was over 0.97. Therefore, the groundwater prediction performance of LSTM was remarkably high. If the groundwater level is simulated on the assumption that the future withdrawal amount is reduced by 1/3 of the current groundwater withdrawal, the range of the maximum rise of the groundwater level would be 0.06–0.13 m compared to the current condition. In addition, assuming that no groundwater is taken, the range of the maximum increase in the groundwater level would be 0.11–0.38 m more than the current condition. Therefore, the effect of groundwater withdrawal on the groundwater level in this area was exceedingly small. The method and results can be used to develop new groundwater withdrawal sources for the redistribution of groundwater withdrawals.


2019 ◽  
Vol 31 (6) ◽  
pp. 1085-1113 ◽  
Author(s):  
Po-He Tseng ◽  
Núria Armengol Urpi ◽  
Mikhail Lebedev ◽  
Miguel Nicolelis

Although many real-time neural decoding algorithms have been proposed for brain-machine interface (BMI) applications over the years, an optimal, consensual approach remains elusive. Recent advances in deep learning algorithms provide new opportunities for improving the design of BMI decoders, including the use of recurrent artificial neural networks to decode neuronal ensemble activity in real time. Here, we developed a long-short term memory (LSTM) decoder for extracting movement kinematics from the activity of large ( N = 134–402) populations of neurons, sampled simultaneously from multiple cortical areas, in rhesus monkeys performing motor tasks. Recorded regions included primary motor, dorsal premotor, supplementary motor, and primary somatosensory cortical areas. The LSTM's capacity to retain information for extended periods of time enabled accurate decoding for tasks that required both movements and periods of immobility. Our LSTM algorithm significantly outperformed the state-of-the-art unscented Kalman filter when applied to three tasks: center-out arm reaching, bimanual reaching, and bipedal walking on a treadmill. Notably, LSTM units exhibited a variety of well-known physiological features of cortical neuronal activity, such as directional tuning and neuronal dynamics across task epochs. LSTM modeled several key physiological attributes of cortical circuits involved in motor tasks. These findings suggest that LSTM-based approaches could yield a better algorithm strategy for neuroprostheses that employ BMIs to restore movement in severely disabled patients.


Author(s):  
Dejiang Kong ◽  
Fei Wu

The widely use of positioning technology has made mining the movements of people feasible and plenty of trajectory data have been accumulated. How to efficiently leverage these data for location prediction has become an increasingly popular research topic as it is fundamental to location-based services (LBS). The existing methods often focus either on long time (days or months) visit prediction (i.e., the recommendation of point of interest) or on real time location prediction (i.e., trajectory prediction). In this paper, we are interested in the location prediction problem in a weak real time condition and aim to predict users' movement in next minutes or hours. We propose a Spatial-Temporal Long-Short Term Memory (ST-LSTM) model which naturally combines spatial-temporal influence into LSTM to mitigate the problem of data sparsity. Further, we employ a hierarchical extension of the proposed ST-LSTM (HST-LSTM) in an encoder-decoder manner which models the contextual historic visit information in order to boost the prediction performance. The proposed HST-LSTM is evaluated on a real world trajectory data set and the experimental results demonstrate the effectiveness of the proposed model.


2020 ◽  
Vol 196 ◽  
pp. 02007
Author(s):  
Vladimir Mochalov ◽  
Anastasia Mochalova

In this paper, the previously obtained results on recognition of ionograms using deep learning are expanded to predict the parameters of the ionosphere. After the ionospheric parameters have been identified on the ionogram using deep learning in real time, we can predict the parameters for some time ahead on the basis of the new data obtained Examples of predicting the ionosphere parameters using an artificial recurrent neural network architecture long short-term memory are given. The place of the block for predicting the parameters of the ionosphere in the system for analyzing ionospheric data using deep learning methods is shown.


Complexity ◽  
2021 ◽  
Vol 2021 ◽  
pp. 1-9
Author(s):  
Abdullah Alharbi ◽  
Wael Alosaimi ◽  
Radhya Sahal ◽  
Hager Saleh

Low heart rate causes a risk of death, heart disease, and cardiovascular diseases. Therefore, monitoring the heart rate is critical because of the heart’s function to discover its irregularity to detect the health problems early. Rapid technological advancement (e.g., artificial intelligence and stream processing technologies) allows healthcare sectors to consolidate and analyze massive health-based data to discover risks by making more accurate predictions. Therefore, this work proposes a real-time prediction system for heart rate, which helps the medical care providers and patients avoid heart rate risk in real time. The proposed system consists of two phases, namely, an offline phase and an online phase. The offline phase targets developing the model using different forecasting techniques to find the lowest root mean square error. The heart rate time-series dataset is extracted from Medical Information Mart for Intensive Care (MIMIC-II). Recurrent neural network (RNN), long short-term memory (LSTM), gated recurrent units (GRU), and bidirectional long short-term memory (BI-LSTM) are applied to heart rate time series. For the online phase, Apache Kafka and Apache Spark have been used to predict the heart rate in advance based on the best developed model. According to the experimental results, the GRU with three layers has recorded the best performance. Consequently, GRU with three layers has been used to predict heart rate 5 minutes in advance.


Sign in / Sign up

Export Citation Format

Share Document