data transforming
Recently Published Documents


TOTAL DOCUMENTS

30
(FIVE YEARS 14)

H-INDEX

5
(FIVE YEARS 1)

Author(s):  
Prismahardi Aji Riyantoko ◽  
Tresna Maulana Fahrudin ◽  
Kartika Maulida Hindrayani ◽  
Amri Muhaimin ◽  
Trimono

Time series is one of method to forecasting the data. The ACEA company has competition with opened the data in the Water Availability and uses the data to forecast. The dataset namely, Aquifers-Petrignano in Italy in water resources field has five parameters e.g. rainfall, temperature, depth to groundwater, drainage volume, and river hydrometry. In our research will be forecast the depth to groundwater data using univariate and multivariate approach of time series using Prophet Method. Prophet method is one of library which develop by Facebook team. We also use the other approach to making the data clean, or the data ready to forecast. We use handle missing data, transforming, differencing, decomposition time series, determine lag, stationary approach, and Augmented Dickey-Fuller (ADF). The all approach will be uses to make sure that the data not appearing the problem while we tried to forecast. In the other describe, we already get the results using univariate and multivariate Prophet method. The multivariate approach has presented the value of MAE 0.82 and RMSE 0.99, it’s better than while we forecast using univariate Prophet.


2021 ◽  
Author(s):  
Lisa Bastarache ◽  
Jeffrey S. Brown ◽  
James J. Cimino ◽  
David A. Dorr ◽  
Peter J. Embi ◽  
...  

2021 ◽  
Vol 10 (9) ◽  
pp. 619
Author(s):  
João Monteiro ◽  
Bruno Martins ◽  
Miguel Costa ◽  
João M. Pires

Datasets collecting demographic and socio-economic statistics are widely available. Still, the data are often only released for highly aggregated geospatial areas, which can mask important local hotspots. When conducting spatial analysis, one often needs to disaggregate the source data, transforming the statistics reported for a set of source zones into values for a set of target zones, with a different geometry and a higher spatial resolution. This article reports on a novel dasymetric disaggregation method that uses encoder–decoder convolutional neural networks, similar to those adopted in image segmentation tasks, to combine different types of ancillary data. Model training constitutes a particular challenge. This is due to the fact that disaggregation tasks are ill-posed and do not entail the direct use of supervision signals in the form of training instances mapping low-resolution to high-resolution counts. We propose to address this problem through self-training. Our method iteratively refines initial estimates produced by disaggregation heuristics and training models with the estimates from previous iterations together with relevant regularization strategies. We conducted experiments related to the disaggregation of different variables collected for Continental Portugal into a raster grid with a resolution of 200 m. Results show that the proposed approach outperforms common alternative methods, including approaches that use other types of regression models to infer the dasymetric weights.


2021 ◽  
Vol 121 (9) ◽  
pp. A46
Author(s):  
H. Pinsky ◽  
B. Jordan ◽  
C. Anselmo ◽  
S. Kaufman ◽  
J. Gibbons ◽  
...  

2021 ◽  
Vol 2 (3) ◽  
pp. 261
Author(s):  
Fredryc Joshua Pa'o ◽  
Hendry Hendry

This study uses a classification system in managing its data. In classification there are several methods provided, one of which is the decision tree method with the C4.5 algorithm this method means a decision tree where the structure is the same as a flowchart where each node signifies an attribute test, each branch presents the test results and the leaf node represents the class or class distribution. The data used is the data of Lake Poso Tourism visitors from 2009 to 2020, then the method used in this study is divided into several stages, namely the data being studied, analyzing the data, transforming data and designing a decision tree with the C4.5 algorithm. The results achieved from this study are that the number of visitors more than 28,984 has a description of "Much" which is dominated by local tourists, while the value with the name "Less" is in foreign tourists. This is one of the important points in determining the right strategy for developing tourism in Lake Poso.


2021 ◽  
Vol 13 (3) ◽  
pp. 1-15
Author(s):  
Rada Chirkova ◽  
Jon Doyle ◽  
Juan Reutter

Assessing and improving the quality of data are fundamental challenges in Big-Data applications. These challenges have given rise to numerous solutions targeting transformation, integration, and cleaning of data. However, while schema design, data cleaning, and data migration are nowadays reasonably well understood in isolation, not much attention has been given to the interplay between standalone tools in these areas. In this article, we focus on the problem of determining whether the available data-transforming procedures can be used together to bring about the desired quality characteristics of the data in business or analytics processes. For example, to help an organization avoid building a data-quality solution from scratch when facing a new analytics task, we ask whether the data quality can be improved by reusing the tools that are already available, and if so, which tools to apply, and in which order, all without presuming knowledge of the internals of the tools, which may be external or proprietary. Toward addressing this problem, we conduct a formal study in which individual data cleaning, data migration, or other data-transforming tools are abstracted as black-box procedures with only some of the properties exposed, such as their applicability requirements, the parts of the data that the procedure modifies, and the conditions that the data satisfy once the procedure has been applied. As a proof of concept, we provide foundational results on sequential applications of procedures abstracted in this way, to achieve prespecified data-quality objectives, for the use case of relational data and for procedures described by standard relational constraints. We show that, while reasoning in this framework may be computationally infeasible in general, there exist well-behaved cases in which these foundational results can be applied in practice for achieving desired data-quality results on Big Data.


2020 ◽  
Vol 9 (1) ◽  
Author(s):  
Regina Raboin

Key themes in Dickens’ novel, transformation and resurrection, darkness and light, and social justice are firmly connected to the work being done in data. Data librarians can make a difference in times like these: resurrecting data, transforming how students, researchers, or the public think about and use data; unearthing and bringing to light historical data that will give context and meaning to an issue; and that accessible data can help address, and perhaps solve, social justice issues.


2020 ◽  
Vol 29 (3) ◽  
pp. 659-667 ◽  
Author(s):  
Hyungsuk Tak ◽  
Kisung You ◽  
Sujit K. Ghosh ◽  
Bingyue Su ◽  
Joseph Kelly
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document