SQL Scorecard for Improved Stability and Performance of Data Warehouses

2016 ◽  
Vol 4 (3) ◽  
pp. 22-37 ◽  
Author(s):  
Nayem Rahman

Scorecard-based measurement techniques are used by organizations to measure the performance of their business operations. A scorecard approach could be applied to a database system to measure performance of SQL (Structured Query Language) being executed and the extent of resources being used by SQL. In a large data warehouse, thousands of jobs run daily via batch cycles to refresh different subject areas. Simultaneously, thousands of queries by business intelligence tools and ad-hoc queries are being executed twenty-four by seven. There needs to be a controlling mechanism to make sure these batch jobs and queries are efficient and do not consume database systems resources more than optimal. The authors propose measurement of SQL query performance via a scorecard tool. The motivation behind using a scorecard tool is to make sure that the resource consumption of SQL queries is predictable and the database system environment is stable. The experimental results show that queries that pass scorecard evaluation criteria tend to utilize optimal level of database systems computing resources. These queries also show improved parallel efficiency (PE) in using computing resources (CPU, I/O and spool space) that demonstrate the usefulness of SQL scorecard.

2021 ◽  
Vol 8 (1) ◽  
Author(s):  
Peter Baumann ◽  
Dimitar Misev ◽  
Vlad Merticariu ◽  
Bang Pham Huu

AbstractMulti-dimensional arrays (also known as raster data or gridded data) play a key role in many, if not all science and engineering domains where they typically represent spatio-temporal sensor, image, simulation output, or statistics “datacubes”. As classic database technology does not support arrays adequately, such data today are maintained mostly in silo solutions, with architectures that tend to erode and not keep up with the increasing requirements on performance and service quality. Array Database systems attempt to close this gap by providing declarative query support for flexible ad-hoc analytics on large n-D arrays, similar to what SQL offers on set-oriented data, XQuery on hierarchical data, and SPARQL and CIPHER on graph data. Today, Petascale Array Database installations exist, employing massive parallelism and distributed processing. Hence, questions arise about technology and standards available, usability, and overall maturity. Several papers have compared models and formalisms, and benchmarks have been undertaken as well, typically comparing two systems against each other. While each of these represent valuable research to the best of our knowledge there is no comprehensive survey combining model, query language, architecture, and practical usability, and performance aspects. The size of this comparison differentiates our study as well with 19 systems compared, four benchmarked to an extent and depth clearly exceeding previous papers in the field; for example, subsetting tests were designed in a way that systems cannot be tuned to specifically these queries. It is hoped that this gives a representative overview to all who want to immerse into the field as well as a clear guidance to those who need to choose the best suited datacube tool for their application. This article presents results of the Research Data Alliance (RDA) Array Database Assessment Working Group (ADA:WG), a subgroup of the Big Data Interest Group. It has elicited the state of the art in Array Databases, technically supported by IEEE GRSS and CODATA Germany, to answer the question: how can data scientists and engineers benefit from Array Database technology? As it turns out, Array Databases can offer significant advantages in terms of flexibility, functionality, extensibility, as well as performance and scalability—in total, the database approach of offering “datacubes” analysis-ready heralds a new level of service quality. Investigation shows that there is a lively ecosystem of technology with increasing uptake, and proven array analytics standards are in place. Consequently, such approaches have to be considered a serious option for datacube services in science, engineering and beyond. Tools, though, vary greatly in functionality and performance as it turns out.


Author(s):  
Andreas Meier ◽  
Günter Schindler ◽  
Nicolas Werro

In practice, information systems are based on very large data collections mostly stored in relational databases. As a result of information overload, it has become increasingly difficult to analyze huge amounts of data and to generate appropriate management decisions. Furthermore, data are often imprecise because they do not accurately represent the world or because they are themselves imperfect. For these reasons, a context model with fuzzy classes is proposed to extend relational database systems. More precisely, fuzzy classes and linguistic variables and terms, together with appropriate membership functions, are added to the database schema. The fuzzy classification query language (fCQL) allows the user to formulate unsharp queries that are then transformed into appropriate SQL statements using the fCQL toolkit so that no migration of the raw data is needed. In addition to the context model with fuzzy classes, fCQL and its implementation are presented here, illustrated by concrete examples.


Author(s):  
Jody Clements ◽  
Tom Dolafi ◽  
Lowell Umayam ◽  
Nicole L. Neubarth ◽  
Stuart Berg ◽  
...  

AbstractDue to technological advances in electron microscopy (EM) and deep learning, it is now practical to reconstruct a connectome, a description of neurons and the connections between them, for significant volumes of neural tissue. The limited scope of past reconstructions meant they were primarily used by domain experts, and performance was not a serious problem. But the new reconstructions, of common laboratory creatures such as the fruit fly Drosophila melanogaster, upend these assumptions. These natural neural networks now contain tens of thousands of neurons and tens of millions of connections between them, with yet larger reconstructions pending, and are of interest to a large community of non-specialists. This requires new tools that are easy to use and efficiently handle large data. We introduce neuPrint to address these data analysis challenges. neuPrint is a database and analysis ecosystem that organizes connectome data in a manner conducive to biological discovery. In particular, we propose a data model that allows users to access the connectome at different levels of abstraction primarily through a graph database, neo4j, and its powerfully expressive query language Cypher. neuPrint is compatible with modern connectome reconstruction workflows, providing tools for assessing reconstruction quality, and offering both batch and incremental updates to match modern connectome reconstruction flows. Finally, we introduce a web interface and programmer API that targets a diverse user skill set. We demonstrate the effectiveness and efficiency of neuPrint through example database queries.


2021 ◽  
Author(s):  
Timo Kersten ◽  
Viktor Leis ◽  
Thomas Neumann

AbstractAlthough compiling queries to efficient machine code has become a common approach for query execution, a number of newly created database system projects still refrain from using compilation. It is sometimes claimed that the intricacies of code generation make compilation-based engines too complex. Also, a major barrier for adoption, especially for interactive ad hoc queries, is long compilation time. In this paper, we examine all stages of compiling query execution engines and show how to reduce compilation overhead. We incorporate the lessons learned from a decade of generating code in HyPer into a design that manages complexity and yields high speed. First, we introduce a code generation framework that establishes abstractions to manage complexity, yet generates code in a single fast pass. Second, we present a program representation whose data structures are tuned to support fast code generation and compilation. Third, we introduce a new compiler backend that is optimized for minimal compile time, and simultaneously, yields superior execution performance to competing approaches, e.g., Volcano-style or bytecode interpretation. We implemented these optimizations in our database system Umbra to show that it is possible to unite fast compilation and fast execution. Indeed, Umbra achieves unprecedentedly low query latencies. On small data sets, it is even faster than interpreter engines like DuckDB and PostgreSQL. At the same time, on large data sets, its throughput is on par with the state-of-the-art compiling system HyPer.


Sensors ◽  
2021 ◽  
Vol 21 (4) ◽  
pp. 1342
Author(s):  
Borja Nogales ◽  
Miguel Silva ◽  
Ivan Vidal ◽  
Miguel Luís ◽  
Francisco Valera ◽  
...  

5G communications have become an enabler for the creation of new and more complex networking scenarios, bringing together different vertical ecosystems. Such behavior has been fostered by the network function virtualization (NFV) concept, where the orchestration and virtualization capabilities allow the possibility of dynamically supplying network resources according to its needs. Nevertheless, the integration and performance of heterogeneous network environments, each one supported by a different provider, and with specific characteristics and requirements, in a single NFV framework is not straightforward. In this work we propose an NFV-based framework capable of supporting the flexible, cost-effective deployment of vertical services, through the integration of two distinguished mobile environments and their networks: small sized unmanned aerial vehicles (SUAVs), supporting a flying ad hoc network (FANET) and vehicles, promoting a vehicular ad hoc network (VANET). In this context, a use case involving the public safety vertical will be used as an illustrative example to showcase the potential of this framework. This work also includes the technical implementation details of the framework proposed, allowing to analyse and discuss the delays on the network services deployment process. The results show that the deployment times can be significantly reduced through a distributed VNF configuration function based on the publish–subscribe model.


2014 ◽  
Vol 14 (2) ◽  
pp. 238-251 ◽  
Author(s):  
Samuel Nana Yaw Simpson

Purpose – This study aims to examine the structure, attributes, and performance of boards of directors of state-owned enterprises (SOEs) within the broader context of public sector governance. This is informed by the less attention given to the concept among public sector organizations despite efforts to make state enterprises more effective and efficient, especially in developing and middle income countries. Design/methodology/approach – Data was collected through questionnaires self-administered in 2010 to all 25 SOEs in Accra, Ghana, out of the 29 nationwide. Some key officials were interviewed and documentary evidence analyzed to achieve triangulation of data and results. Findings – Results show that state-owned enterprises have boards and comply with the minimal governance issues outlined the legal frameworks establishing them. However, they exhibit significant weaknesses in the areas of board performance evaluation, criteria for board appointment, the balance of executive directors and non-executive directors, and other board characteristics, indicating a departure from general practices. Practical implications – Findings suggest the need for a tailored corporate governance framework or code for state-owned enterprises in developing countries. Originality/value – Compared to the literature, this study provides insight on boards from the perspective of state enterprises in ensuring good corporate governance, particularly in the context of a middle income country (Ghana).


2018 ◽  
Vol 60 (4) ◽  
pp. 335-354 ◽  
Author(s):  
Marco Botta

This study investigates the existence of an optimal capital structure for small and medium enterprise (SME) hotels through the analysis of the relationship between financing decisions and financial performance in a large sample of Italian hotel SMEs. The results show that hotel SMEs face an optimal capital structure that allows them to maximize returns to investors, while instead having both too little and too much debt reduces their financial performance. This notwithstanding, we show that hotel SMEs are not particularly concerned with optimizing their capital structure, and their funding behavior is deeply connected with the availability of internally available funds, a typical pecking order behavior, and they result extremely slow in converging toward their optimal level of leverage so that they could improve their performance by adopting a more sophisticated financial strategy.


2018 ◽  
Vol 14 (11) ◽  
pp. 155014771881505 ◽  
Author(s):  
Ishtiaq Wahid ◽  
Ata Ul Aziz Ikram ◽  
Masood Ahmad ◽  
Fasee Ullah

With resource constraint’s distributed architecture and dynamic topology, network issues such as congestion, latency, power awareness, mobility, and other quality of service issues need to be addressed by optimizing the routing protocols. As a result, a number of routing protocols have been proposed. Routing protocols have trade-offs in performance parameters and their performance varies with the underlying mobility model. For designing an improved vehicular ad hoc network, three components of the network are to be focused: routing protocols, mobility models, and performance metrics. This article describes the relationship of these components, trade-offs in performance, and proposes a supervisory protocol, which monitors the scenario and detects the realistic mobility model through analysis of the microscopic features of the mobility model. An analytical model is used to determine the best protocol for a particular mobility model. The supervisory protocol then selects the best routing protocol for the mobility model of the current operational environment. For this, EstiNet 8.1 Simulator is used to validate the proposed scheme and compare its performance with existing schemes. Simulation results of the proposed scheme show the consistency in the performance of network throughout its operation.


Author(s):  
Hendra Galuh Febrianto ◽  
Amalia Indah Fitriana

ABSTRACT In the banking world of soundness, banks are very important for the formation of trust. Trust and loyalty to banks is a very helpful factor and makes it easier for bank management to develop good business strategies. Bank Soundness Levels are results issued by banks which are carried out on bank risk and performance (Bank Indonesia Regulation Number: 13/1 / PBI / 2011). If more than conventional banking with Islamic banking, conventional banking finance is better than Islamic banking. This is blessed with poor sharia banking (corporate governance) management. In order to be able to carry out its functions properly, banks must have sufficient capital, ensure the quality of their assets properly, be well managed and managed based on the principle of prudence, generate sufficient profits to maintain an increase, and support liquidity so that it can be adjusted to their needs. Therefore banks are required to be able to achieve and maintain a good and optimal level of performance, because the level of bank performance can increase the level of trust and loyalty needed by the wider community to use the products, services and financial activities of the bank. The purpose of this study is for advanced financial research with analysis of Risk Profiles (Risk Profiles), Good Corporate Governance (GCG), Profitability (Income), and Capital (Capital) which is hereinafter abbreviated as RGEC with the final aim of research for the needs of Sharia banking management in accordance with the latest Bank Indonesia and OJK regulations. This type of research uses descriptive research proposed in the RGEC analysis (Risk Profile, Good Corporate Governance, Income, and Capital) at Islamic Banks in Indonesia. from 2013 to 2017. Keywords: Risk Profile, Good Corporate Governance, Income, Capital, Bank Soundness   ABSTRAK Dalam dunia perbankan tingkat kesehatan bank sangat penting bagi pembentukan kepercayaan. Kepercayaan dan loyalitas nasabah terhadap bank merupakan faktor yang sangat membantu dan mempermudah pihak manajemen bank untuk menyusun strategi bisnis yang baik. Tingkat Kesehatan Bank adalah hasil penilaian kondisi bank yang dilakukan terhadap risiko dan kinerja bank (Peraturan Bank Indonesia Nomor: 13/1/PBI/2011). Jika dibanding antara perbankan konvensional dengan perbankan syariah, kinerja keuangan perbankan konvensional lebih baik daripada perbankan syariah. Hal ini dikarena tatakelola (good corporate governance) perbankan syariah yang masih buruk. Agar dapat menjalankan fungsinya dengan baik, bank harus mempunyai modal yang cukup, menjaga kualitas asetnya dengan baik, dikelola dengan baik dan dioperasikan berdasarkan prinsip kehati-hatian, menghasilkan keuntungan yang cukup untuk mempertahankan kelangsungan usahanya, serta memelihara likuiditasnya sehingga dapat memenuhi kewajibannya. Oleh karena itu bank dituntut untuk bisa mencapai dan mempertahankan tingkat kinerja yang baik dan optimal, karena tingkat kinerja bank yang baik dapat meningkatkan kepercayaan dan loyalitas nasabah maupun masyarakat luas untuk menggunakan produk, jasa dan aktivitas keuangan dari bank tersebut. Tujuan penelitian ini adalah untuk menilai tingkat kesehatan keuangan dengan analisis Profil Risiko (Risk Profile), Good Corporate Governance (GCG), Rentabilitas (Earnings), dan Permodalan (Capital) yang selanjutnya disingkat RGEC dengan tujuan akhir merekomendasikan kebijakan untuk memperbaiki manajemen perbankan Syariah yang sesuai peraturan Bank Indonesia dan OJK yang terbaru. Jenis penelitian ini menggunakan penelitian deskriptif yang berfokus pada analisis RGEC (Risk Profile, Good Corporate Governance, Earnings, and Capital) pada Bank Syariah di Indonesia. dari tahun 2013 sampai 2017. Kata kunci: Risk Profile, Good Corporate Governance, Earnings, Capital, Tingkat Kesehatan Bank


Sign in / Sign up

Export Citation Format

Share Document