A Single Pass Algorithm for Discovering Significant Intervals in Time-Series Data

2008 ◽  
pp. 3272-3284
Author(s):  
Sagar Savla ◽  
Sharma Chakravarthy

Sensor-based applications, such as smart homes, require prediction of event occurrences for automating the environment using time-series data collected over a period of time. In these applications, it is important to predict events in tight and accurate intervals to effectively automate the application. This article deals with the discovery of significant intervals from time-series data. Although there is a considerable body of work on sequential mining of transactional data, most of them deal with time-point data and make several passes over the entire data set in order to discover frequently occurring patterns/events. We propose an approach in which significant intervals representing intrinsic nature of data are discovered in a single pass. In our approach, time-series data is folded over a periodicity (day, week, etc.) in which the intervals are formed. Significant intervals are discovered from this interval data that satisfy the criteria of minimum confidence and maximum interval length specified by the user. Both compression and working with intervals contribute towards improving the efficiency of the algorithm. In this article, we present a new single-pass algorithm for detecting significant intervals; discuss its characteristics, advantages, and disadvantages; and analyze it. Finally, we compare the performance of our algorithm with previously developed level-wise and SQL-based algorithms for significant interval discovery (SID).

AI ◽  
2021 ◽  
Vol 2 (1) ◽  
pp. 48-70
Author(s):  
Wei Ming Tan ◽  
T. Hui Teo

Prognostic techniques attempt to predict the Remaining Useful Life (RUL) of a subsystem or a component. Such techniques often use sensor data which are periodically measured and recorded into a time series data set. Such multivariate data sets form complex and non-linear inter-dependencies through recorded time steps and between sensors. Many current existing algorithms for prognostic purposes starts to explore Deep Neural Network (DNN) and its effectiveness in the field. Although Deep Learning (DL) techniques outperform the traditional prognostic algorithms, the networks are generally complex to deploy or train. This paper proposes a Multi-variable Time Series (MTS) focused approach to prognostics that implements a lightweight Convolutional Neural Network (CNN) with attention mechanism. The convolution filters work to extract the abstract temporal patterns from the multiple time series, while the attention mechanisms review the information across the time axis and select the relevant information. The results suggest that the proposed method not only produces a superior accuracy of RUL estimation but it also trains many folds faster than the reported works. The superiority of deploying the network is also demonstrated on a lightweight hardware platform by not just being much compact, but also more efficient for the resource restricted environment.


MAUSAM ◽  
2021 ◽  
Vol 68 (2) ◽  
pp. 349-356
Author(s):  
J. HAZARIKA ◽  
B. PATHAK ◽  
A. N. PATOWARY

Perceptive the rainfall pattern is tough for the solution of several regional environmental issues of water resources management, with implications for agriculture, climate change, and natural calamity such as floods and droughts. Statistical computing, modeling and forecasting data are key instruments for studying these patterns. The study of time series analysis and forecasting has become a major tool in different applications in hydrology and environmental fields. Among the most effective approaches for analyzing time series data is the ARIMA (Autoregressive Integrated Moving Average) model introduced by Box and Jenkins. In this study, an attempt has been made to use Box-Jenkins methodology to build ARIMA model for monthly rainfall data taken from Dibrugarh for the period of 1980- 2014 with a total of 420 points.  We investigated and found that ARIMA (0, 0, 0) (0, 1, 1)12 model is suitable for the given data set. As such this model can be used to forecast the pattern of monthly rainfall for the upcoming years, which can help the decision makers to establish priorities in terms of agricultural, flood, water demand management etc.  


Sensors ◽  
2019 ◽  
Vol 19 (12) ◽  
pp. 2812 ◽  
Author(s):  
Jing Yang ◽  
Yizhong Sun ◽  
Bowen Shang ◽  
Lei Wang ◽  
Jie Zhu

With the availability of large geospatial datasets, the study of collective human mobility spatiotemporal patterns provides a new way to explore urban spatial environments from the perspective of residents. In this paper, we constructed a classification model for mobility patterns that is suitable for taxi OD (Origin-Destination) point data, and it is comprised of three parts. First, a new aggregate unit, which uses a road intersection as the constraint condition, is designed for the analysis of the taxi OD point data. Second, the time series similarity measurement is improved by adding a normalization procedure and time windows to address the particular characteristics of the taxi time series data. Finally, the DBSCAN algorithm is used to classify the time series into different mobility patterns based on a proximity index that is calculated using the improved similarity measurement. In addition, we used the random forest algorithm to establish a correlation model between the mobility patterns and the regional functional characteristics. Based on the taxi OD point data from Nanjing, we delimited seven mobility patterns and illustrated that the regional functions have obvious driving effects on these mobility patterns. These findings are applicable to urban planning, traffic management and planning, and land use analyses in the future.


Author(s):  
T. Warren Liao

In this chapter, we present genetic algorithm (GA) based methods developed for clustering univariate time series with equal or unequal length as an exploratory step of data mining. These methods basically implement the k-medoids algorithm. Each chromosome encodes in binary the data objects serving as the k-medoids. To compare their performance, both fixed-parameter and adaptive GAs were used. We first employed the synthetic control chart data set to investigate the performance of three fitness functions, two distance measures, and other GA parameters such as population size, crossover rate, and mutation rate. Two more sets of time series with or without known number of clusters were also experimented: one is the cylinder-bell-funnel data and the other is the novel battle simulation data. The clustering results are presented and discussed.


2004 ◽  
Vol 91 (3-4) ◽  
pp. 332-344 ◽  
Author(s):  
Jin Chen ◽  
Per. Jönsson ◽  
Masayuki Tamura ◽  
Zhihui Gu ◽  
Bunkei Matsushita ◽  
...  

2020 ◽  
Vol 12 (01) ◽  
pp. 2050001
Author(s):  
Yadigar N. Imamverdiyev ◽  
Fargana J. Abdullayeva

In this paper, a fault prediction method for oil well equipment based on the analysis of time series data obtained from multiple sensors is proposed. The proposed method is based on deep learning (DL). For this purpose, comparative analysis of single-layer long short-term memory (LSTM) with the convolutional neural network (CNN) and stacked LSTM methods is provided. To demonstrate the efficacy of the proposed method, some experiments are conducted on the real data set obtained from eight sensors installed in oil wells. In this paper, compared to the single-layer LSTM model, the CNN and stacked LSTM predicted the faulty time series with a minimal loss.


Author(s):  
Yoshiyuki Matsumoto ◽  
◽  
Junzo Watada ◽  

Rough sets theory was proposed by Z. Pawlak in 1982. This theory enables us to mine knowledge granules through a decision rule from a database, a web base, a set and so on. We can apply the decision rule to reason, estimate, evaluate, or forecast unknown objects. In this paper, the rough set model is used to analyze of time series data of tick-wise price fluctuation, where knowledge granules are mined from the data set of tick-wise price fluctuations.


Sign in / Sign up

Export Citation Format

Share Document