Adaptive data reduction for large-scale transaction data

2008 ◽  
Vol 188 (3) ◽  
pp. 910-924 ◽  
Author(s):  
Xiao-Bai Li ◽  
Varghese S. Jacob
2021 ◽  
Vol 33 (12) ◽  
pp. 1795-1802
Author(s):  
Zhiwei Ai ◽  
Juelin Leng ◽  
Fang Xia ◽  
Huawei Wang ◽  
Yi Cao

Author(s):  
Veronika Strnadová-Neeley ◽  
Aydın Buluç ◽  
Jarrod Chapman ◽  
John R. Gilbert ◽  
Joseph Gonzalez ◽  
...  

2012 ◽  
Vol 45 (2) ◽  
pp. 324-328 ◽  
Author(s):  
Jan Ilavsky

Nikais anIgor Pro-based package for correction, calibration and reduction of two-dimensional area-detector data into one-dimensional data (`lineouts'). It is free (although the user needs a paid license forIgor Pro), open source and highly flexible. While typically used for small-angle X-ray scattering (SAXS) data, it can also be used for grazing-incidence SAXS data, wide-angle diffraction data and even small-angle neutron scattering data. It has been widely available to the user community since about 2005, and it is currently used at the SAXS instruments of selected large-scale facilities as their main data reduction package. It is, however, also suitable for desktop instruments when the manufacturer's software is not available or appropriate. Since it is distributed as source code, it can be scrutinized, verified and modified by users to suit their needs.


2014 ◽  
Vol 104 (9) ◽  
pp. 2797-2829 ◽  
Author(s):  
Ing-Haw Cheng ◽  
Sahil Raina ◽  
Wei Xiong

We analyze whether midlevel managers in securitized finance were aware of a large-scale housing bubble and a looming crisis in 2004–2006 using their personal home transaction data. We find that the average person in our sample neither timed the market nor were cautious in their home transactions, and did not exhibit awareness of problems in overall housing markets. Certain groups of securitization agents were particularly aggressive in increasing their exposure to housing during this period, suggesting the need to expand the incentives-based view of the crisis to incorporate a role for beliefs. (JEL D14, D83, E32, E44, G01, G21, R31)


2018 ◽  
Vol 7 (3) ◽  
pp. 56-75 ◽  
Author(s):  
Omdutt Sharma ◽  
Pratiksha Tiwari ◽  
Priti Gupta

This article describes how information technology and internet together infused organizations with huge amount of data. Consequently, accumulating, storing, understanding and analyzing data at a large scale is equally important and complex. Out of this data not all is information data, in order to extract information, one needs to discard redundant, irrelevant and unnecessary data. This article aims to introduce a data reduction technique which will be useful to discard irrelevant data. Here in data-reduction, the authors have used fuzzy-soft set techniques, namely fuzzy-soft information matrixes. Further, they have introduced a new fuzzy-soft information measure of fuzzy-soft matrixes.


2021 ◽  
Author(s):  
◽  
Gijs Jan Molenaar

The preparation for the construction of the Square Kilometre Array, and the introduction of its operational precursors, such as LOFAR and MeerKAT, mark the beginning of an exciting era for astronomy. Impressive new data containing valuable science just waiting for discovery is already being generated, and these devices will produce far more data than has ever been collected before. However, with every new data instrument, the data rates grow to unprecedented quantities of data, requiring novel new data-processing tools. In addition, creating science grade data from the raw data still requires significant expert knowledge for processing this data. The software used is often developed by a scientist who lacks proper training in software development skills, resulting in the software not progressing beyond a prototype stage in quality. In the first chapter, we explore various organisational and technical approaches to address these issues by providing a historical overview of the development of radioastronomy pipelines since the inception of the field in the 1940s. In that, the steps required to create a radio image are investigated. We used the lessons-learned to identify patterns in the challenges experienced, and the solutions created to address these over the years. The second chapter describes the mathematical foundations that are essential for radio imaging. In the third chapter, we discuss the production of the KERN Linux distribution, which is a set of software packages containing most radio astronomy software currently in use. Considerable effort was put into making sure that the contained software installs appropriately, all items next to one other on the same system. Where required and possible, bugs and portability fixes were solved and reported with the upstream maintainers. The KERN project also has a website, and issue tracker, where users can report bugs and maintainers can coordinate the packaging effort and new releases. The software packages can be used inside Docker and Singularity containers, enabling the installation of these packages on a wide variety of platforms. In the fourth and fifth chapters, we discuss methods and frameworks for combining the available data reduction tools into recomposable pipelines and introduce the Kliko specification and software. This framework was created to enable end-user astronomers to chain and containerise operations of software in KERN packages. Next, we discuss the Common Workflow Language (CommonWL), a similar but more advanced and mature pipeline framework invented by bio-informatics scientists. CommonWL is supported by a wide range of tools already; among other schedulers, visualisers and editors. Consequently, when a pipeline is made with CommonWL, it can be deployed and manipulated with a wide range of tools. In the final chapter, we attempt something unconventional, applying a generative adversarial network based on deep learning techniques to perform the task of sky brightness reconstruction. Since deep learning methods often require a large number of training samples, we constructed a CommonWL simulation pipeline for creating dirty images and corresponding sky models. This simulated dataset has been made publicly available as the ASTRODECONV2019 dataset. It is shown that this method is useful to perform the restoration and matches the performance of a single clean cycle. In addition, we incorporated domain knowledge by adding the point spread function to the network and by utilising a custom loss function during training. Although it was not possible to improve the cleaning performance of commonly used existing tools, the computational time performance of the approach looks very promising. We suggest that a smaller scope should be the starting point for further studies and optimising of the training of the neural network could produce the desired results.


2016 ◽  
Vol 172 ◽  
pp. 189-197 ◽  
Author(s):  
Xiang-Jun Shen ◽  
Lei Mu ◽  
Zhen Li ◽  
Hao-Xiang Wu ◽  
Jian-Ping Gou ◽  
...  

2019 ◽  
Vol 145 (10) ◽  
pp. 04019075 ◽  
Author(s):  
M. K. ElSawy ◽  
M. H. El Naggar ◽  
A. B. Cerato ◽  
A. W. Elgamal

Sign in / Sign up

Export Citation Format

Share Document