scholarly journals Uncertainty-Aware Multiple Instance Learning from Large-Scale Long Time Series Data

Author(s):  
Yuansheng Zhu ◽  
Weishi Shi ◽  
Deep Shankar Pandey ◽  
Yang Liu ◽  
Xiaofan Que ◽  
...  
2020 ◽  
Vol 245 ◽  
pp. 07001
Author(s):  
Laura Sargsyan ◽  
Filipe Martins

Large experiments in high energy physics require efficient and scalable monitoring solutions to digest data of the detector control system. Plotting multiple graphs in the slow control system and extracting historical data for long time periods are resource intensive tasks. The proposed solution leverages the new virtualization, data analytics and visualization technologies such as InfluxDB time-series database for faster access large scale data, Grafana to visualize time-series data and an OpenShift container platform to automate build, deployment, and management of application. The monitoring service runs separately from the control system thus reduces a workload on the control system computing resources. As an example, a test version of the new monitoring was applied to the ATLAS Tile Calorimeter using the CERN Cloud Process as a Service platform. Many dashboards in Grafana have been created to monitor and analyse behaviour of the High Voltage distribution system. They visualize not only values measured by the control system, but also run information and analytics data (difference, deviation, etc.). The new monitoring with a feature-rich visualization, filtering possibilities and analytics tools allows to extend detector control and monitoring capabilities and can help experts working on large scale experiments.


Sensors ◽  
2021 ◽  
Vol 21 (13) ◽  
pp. 4466
Author(s):  
Li Guo ◽  
Runze Li ◽  
Bin Jiang

The monitoring of electrical equipment and power grid systems is very essential and important for power transmission and distribution. It has great significances for predicting faults based on monitoring a long sequence in advance, so as to ensure the safe operation of the power system. Many studies such as recurrent neural network (RNN) and long short-term memory (LSTM) network have shown an outstanding ability in increasing the prediction accuracy. However, there still exist some limitations preventing those methods from predicting long time-series sequences in real-world applications. To address these issues, a data-driven method using an improved stacked-Informer network is proposed, and it is used for electrical line trip faults sequence prediction in this paper. This method constructs a stacked-Informer network to extract underlying features of long sequence time-series data well, and combines the gradient centralized (GC) technology with the optimizer to replace the previously used Adam optimizer in the original Informer network. It has a superior generalization ability and faster training efficiency. Data sequences used for the experimental validation are collected from the wind and solar hybrid substation located in Zhangjiakou city, China. The experimental results and concrete analysis prove that the presented method can improve fault sequence prediction accuracy and achieve fast training in real scenarios.


Eos ◽  
2017 ◽  
Vol 98 ◽  
Author(s):  
Toste Tanhua

How measurements from a glider deployed off the coast of Peru are contributing to a much-needed long time-series data set.


Author(s):  
S. Jing ◽  
T. Chao

Abstract. Time series imagery containing high-dimensional temporal features are conducive to improving classification accuracy. With the plenty accumulation of historical images, the inclusion of time series data becomes available to utilize, but it is difficult to avoid missing values caused by cloud cover. Meanwhile, seeking a large amount of training labels for long time series also makes data collection troublesome. In this study, we proposed a semi-supervised convolutional long short-term memory neural network (Semi-LSTM) in long time series which achieves an accurate and automated land cover classification with a small proportion of labels. Three main contributions of this work are summarized as follows: i) the proposed method achieve an excellent classification via a small group of labels in long time series data, and reducing dependence of training labels; ii) it is a robust algorithm in accuracy for the influence of noise, and reduces the requirements of sequential data for cloudless and lossless images; and iii) it makes full advantage of spectral-spatial-temporal features, especially expanding time context information to enhance classification accuracy. Finally, the proposed network is validated on time series imagery from Landsat 8. All quantitative analyses and evaluation indicators of the experimental results demonstrate competitive performance in the suggested modes.


Author(s):  
Joyce Ann Guzik

The NASA Kepler and follow-on K2 mission (2009–2018) left a legacy of data and discoveries, finding thousands of exoplanets, and also obtaining high-precision long time-series data for hundreds of thousands of stars, including many types of pulsating variables. Here we highlight a few of the ongoing discoveries from Kepler data on δ Scuti pulsating variables, which are core hydrogen-burning stars of about twice the mass of the Sun. We discuss many unsolved problems surrounding the properties of the variability in these stars, and the progress enabled by Kepler data in using pulsations to infer their interior structure, a field of research known as asteroseismology.


Fractals ◽  
2010 ◽  
Vol 18 (03) ◽  
pp. 301-307 ◽  
Author(s):  
ALFONSO GARMENDIA ◽  
LUIS GARMENDIA ◽  
ADELA SALVADOR

Can the fractal dimension of fluctuations in population size be used to estimate extinction risk? The problem with estimating this fractal dimension is that the lengths of the time series are usually too short for conclusive results. This study answered this question with long time series data obtained from an iterative competition model. This model produces competitive extinction at different perturbation intensities for two different germination strategies: germination of all seeds vs. dormancy in half the seeds. This provided long time series of 900 years and different extinction risks. The results support the hypothesis for the effectiveness of the Hurst coefficient for estimating extinction risk.


2007 ◽  
pp. 88
Author(s):  
Wataru Suzuki ◽  
Yanfei Zhou

This article represents the first step in filling a large gap in knowledge concerning why Public Assistance (PA) use recently rose so fast in Japan. Specifically, we try to address this problem not only by performing a Blanchard and Quah decomposition on long-term monthly time series data (1960:04-2006:10), but also by estimating prefecturelevel longitudinal data. Two interesting findings emerge from the time series analysis. The first is that permanent shock imposes a continuously positive impact on the PA rate and is the main driving factor behind the recent increase in welfare use. The second finding is that the impact of temporary shock will last for a long time. The rate of the use of welfare is quite rigid because even if the PA rate rises due to temporary shocks, it takes about 8 or 9 years for it to regain its normal level. On the other hand, estimations of prefecture-level longitudinal data indicate that the Financial Capability Index (FCI) of the local government2 and minimum wage both impose negative effects on the PA rate. We also find that the rapid aging of Japan's population presents a permanent shock in practice, which makes it the most prominent contribution to surging welfare use.


2021 ◽  
Author(s):  
Sadnan Al Manir ◽  
Justin Niestroy ◽  
Maxwell Adam Levinson ◽  
Timothy Clark

Introduction: Transparency of computation is a requirement for assessing the validity of computed results and research claims based upon them; and it is essential for access to, assessment, and reuse of computational components. These components may be subject to methodological or other challenges over time. While reference to archived software and/or data is increasingly common in publications, a single machine-interpretable, integrative representation of how results were derived, that supports defeasible reasoning, has been absent. Methods: We developed the Evidence Graph Ontology, EVI, in OWL 2, with a set of inference rules, to provide deep representations of supporting and challenging evidence for computations, services, software, data, and results, across arbitrarily deep networks of computations, in connected or fully distinct processes. EVI integrates FAIR practices on data and software, with important concepts from provenance models, and argumentation theory. It extends PROV for additional expressiveness, with support for defeasible reasoning. EVI treats any com- putational result or component of evidence as a defeasible assertion, supported by a DAG of the computations, software, data, and agents that produced it. Results: We have successfully deployed EVI for very-large-scale predictive analytics on clinical time-series data. Every result may reference its own evidence graph as metadata, which can be extended when subsequent computations are executed. Discussion: Evidence graphs support transparency and defeasible reasoning on results. They are first-class computational objects, and reference the datasets and software from which they are derived. They support fully transparent computation, with challenge and support propagation. The EVI approach may be extended to include instruments, animal models, and critical experimental reagents.


2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Jing Zhao ◽  
Shubo Liu ◽  
Xingxing Xiong ◽  
Zhaohui Cai

Privacy protection is one of the major obstacles for data sharing. Time-series data have the characteristics of autocorrelation, continuity, and large scale. Current research on time-series data publication mainly ignores the correlation of time-series data and the lack of privacy protection. In this paper, we study the problem of correlated time-series data publication and propose a sliding window-based autocorrelation time-series data publication algorithm, called SW-ATS. Instead of using global sensitivity in the traditional differential privacy mechanisms, we proposed periodic sensitivity to provide a stronger degree of privacy guarantee. SW-ATS introduces a sliding window mechanism, with the correlation between the noise-adding sequence and the original time-series data guaranteed by sequence indistinguishability, to protect the privacy of the latest data. We prove that SW-ATS satisfies ε-differential privacy. Compared with the state-of-the-art algorithm, SW-ATS is superior in reducing the error rate of MAE which is about 25%, improving the utility of data, and providing stronger privacy protection.


Sign in / Sign up

Export Citation Format

Share Document