Effectively and Efficiently Designing and Querying Parallel Relational Data Warehouses on Heterogeneous Database Clusters

2012 ◽  
Vol 23 (4) ◽  
pp. 17-51 ◽  
Author(s):  
Ladjel Bellatreche ◽  
Alfredo Cuzzocrea ◽  
Soumia Benkrid

In this paper, a comprehensive methodology for designing and querying Parallel Rational Data Warehouses (PRDW) over database clusters, called Fragmentation & Allocation (F&A) is proposed. F&A assumes that cluster nodes are heterogeneous in processing power and storage capacity, contrary to traditional design approaches that assume that cluster nodes are instead homogeneous, and fragmentation and allocation phases are performed in a simultaneous manner. In classical approaches, two different cost models are used to perform fragmentation and allocation, separately, whereas F&A makes use of one cost model that considers fragmentation and allocation parameters simultaneously. Therefore, according to the F&A methodology proposed, the allocation phase/decision is done at fragmentation. At the fragmentation phase, F&A uses two well-known algorithms, namely Hill Climbing (HC) and Genetic Algorithm (GA), which the authors adapt to the main PRDW design problem over heterogeneous database clusters, as these algorithms are capable of taking into account the heterogeneous characteristics of the reference application scenario. At the allocation phase, F&A introduces an innovative matrix-based formalism capable of capturing the interactions among fragments, input queries, and cluster node characteristics, driving the data allocation task accordingly, and a related affinity-based algorithm, called F&A-ALLOC. Finally, their proposal is experimentally assessed and validated against the widely-known data warehouse benchmark APB-1 release II.

Author(s):  
Cyrus Shahabi ◽  
Dimitris Sacharidis ◽  
Mehrdad Jahangiri

Following the constant technological advancements that provide more processing power and storage capacity, scientific applications have emerged as a new field of interest for the database community. Such applications, termed Online Science Applications (OSA), require continuous interaction with datasets of multidimensional nature, mainly for performing statistical analysis. OSA can seriously benefit from the ongoing research for OLAP systems and the pre-calculation of aggregate functions for multidimensional datasets. One of the tools that we see fit for the task in hand is the wavelet transformation. Due to its inherent multi-resolution properties, wavelets can be utilized to provide progressively approximate and eventually fast exact answers to complex queries in the context of Online Science Applications.


Author(s):  
Ladjel Bellatreche ◽  
Kamel Boukhalfa ◽  
Pascal Richard

Horizontal partitioning has evolved significantly in recent years and widely advocated by the academic and industrial communities. Horizontal Partitioning affects positively query performance, database manageability and availability. Two types of horizontal partitioning are supported: primary and referential. Horizontal fragmentation in the context of relational data warehouses is to partition dimension tables by primary fragmentation then fragmenting the fact table by referential fragmentation. This fragmentation can generate a very large number of fragments which may make the maintenance task very complicated. In this paper, we first focus on the evolution of horizontal partitioning in commercial DBMS motivated by decision support applications. Secondly, we give a formalization of the referential fragmentation schema selection problem in the data warehouse and we study its hardness to select an optimal solution. Due to its high complexity, we develop two algorithms: hill climbing and simulated annealing with several variants to select a near optimal partitioning schema. We present ParAdmin, an advisor tool assisting administrators to use primary and referential partitioning during the physical design of their data warehouses. Finally, extensive experimental studies are conducted using the data set of APB1 benchmark to compare the quality the proposed algorithms using a mathematical cost model. Based on these experiments, some recommendations are given to ensure the well use of horizontal partitioning.


English Today ◽  
2008 ◽  
Vol 24 (1) ◽  
pp. 23-27 ◽  
Author(s):  
Michael Rundell

ABSTRACTStarts with an excerpt from Michael Rundell and Penny Stock, The Corpus revolution (ET30, 1992). An update on the rise and rise of electronic language corpora and their impact on dictionaries. How dramatically the world has changed since Penny Stock and I wrote about the ‘Corpus Revolution’ in 1992.At the time, it was not hard to predict that computer processing power and storage capacity would carry on doubling each year. It was already clear, too, that the arrival of the corpus would revolutionize the work of dictionary-makers – hence the title of our articles. These changes were well under way in 1992 and, sixteen years on, their effects are still being felt. In the process, dictionaries have got dramatically better – if by ‘better’ we mean that the description of language they provide corresponds more closely to the way people actually use words when they communicate with one another.


Sensors ◽  
2019 ◽  
Vol 19 (6) ◽  
pp. 1339 ◽  
Author(s):  
Hasan Islam ◽  
Dmitrij Lagutin ◽  
Antti Ylä-Jääski ◽  
Nikos Fotiou ◽  
Andrei Gurtov

The Constrained Application Protocol (CoAP) is a specialized web transfer protocol which is intended to be used for constrained networks and devices. CoAP and its extensions (e.g., CoAP observe and group communication) provide the potential for developing novel applications in the Internet-of-Things (IoT). However, a full-fledged CoAP-based application may require significant computing capability, power, and storage capacity in IoT devices. To address these challenges, we present the design, implementation, and experimentation with the CoAP handler which provides transparent CoAP services through the ICN core network. In addition, we demonstrate how the CoAP traffic over an ICN network can unleash the full potential of the CoAP, shifting both overhead and complexity from the (constrained) endpoints to the ICN network. The experiments prove that the CoAP Handler helps to decrease the required computation complexity, communication overhead, and state management of the CoAP server.


1975 ◽  
Vol 65 (2) ◽  
pp. 359-372 ◽  
Author(s):  
H J Reimers ◽  
D J Allen ◽  
I A Feuerstein ◽  
J F Mustard

Repeated thrombin treatment of washed platelets prepared from rabbits can decrease the serotonin content of the platelets by about 80%. When these platelets are deaggregated they reaccumulate serotonin but their storage capacity for serotonin is reduced by about 60%. If thrombin-pretreated platelets are allowed to equilibrate with a high concentration of serotonin (123 mu M), they release a smaller percentage of their total serotonin upon further thrombin treatment, in comparison with the percentage of serotonin released from control platelets equilibrated with the same concentration of serotonin calculations indicate that in thrombin-treated platelets reequilibrated with serotonin, two-thirds of the serotonin is in the granule compartment and one-third is in the extragranular compartment, presumably the cytoplasm. Analysis of the exchange of serotonin between the suspending fluid and the platelets showed that thrombin treatment does not alter the transport rate of serotonin across the platelet membrane and does not cause increased diffusion of serotonin from the platelets into the suspending fluid. The primary reason for the reduced serotonin accumulation by the thrombin-treated platelets appears to be loss of amine storage granules or of the storage capacity within the granules.


2021 ◽  
Vol 11 (2) ◽  
Author(s):  
Brendan P. Marsh ◽  
Yudan Guo ◽  
Ronen M. Kroeze ◽  
Sarang Gopalakrishnan ◽  
Surya Ganguli ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document