Control-Based Database Tuning Under Dynamic Workloads

Author(s):  
Yi-Cheng Tu ◽  
Gang Ding

Database administration (tuning) is the process of adjusting database configurations in order to accomplish desirable performance goals. This job is performed by human operators called database administrators (DBAs) who are generally well-paid, and are becoming more and more expensive with the increasing complexity and scale of modern databases. There has been considerable effort dedicated to reducing such cost (which often dominates the total ownership cost of missioncritical databases) by making database tuning more automated and transparent to users (Chaudhuri et al, 2004; Chaudhuri and Weikum, 2006). Research in this area seeks ways to automate the hardware deployment, physical database design, parameter configuration, and resource management in such systems. The goal is to achieve acceptable performance on the whole system level without (or with limited) human intervention. According to Weikum et al. (2002), problems in this category can be stated as: workload × configuration (?) ? performance which means that, given the features of the incoming workload to the database, we are to find the right settings for all system knobs such that the performance goals are satisfied The following two are representatives of a series of such tuning problems in different databases: • Problem 1: Maintenance of multi-class servicelevel agreements (SLA) in relational databases. Database service providers usually offer various levels of performance guarantees to requests from different groups of customers. Fulfillment of such guarantees (SLAs) is accomplished by allocating different amounts of system resources to different queries. For example, query response time is negatively related to the amount of memory buffer assigned to that query. We need to dynamically allocate memory to individual queries such that the absolute or relative response times of queries from different users are satisfied. • Problem 2: Load shedding in stream databases. Stream databases are used for processing data generated continuously from sources such as a sensor network. In streaming databases, data processing delay, i.e., the time consumed to process a data point, is the most critical performance metric (Tatbul et al., 2003). The ability to remain within a desired level of delay is significantly hampered under situations of overloading (caused by bursty data arrivals and time-varying unit data processing cost). When overloaded, some data is discarded (i.e., load shedding) in order to keep pace with the incoming load. The system needs to continuously adjust the amount of data to be discarded such that 1) delay is maintained under a desirable level; 2) data is not discarded unnecessarily.

2009 ◽  
pp. 2564-2571
Author(s):  
Yi-Cheng Tu ◽  
Gang Ding

Database administration (tuning) is the process of adjusting database configurations in order to accomplish desirable performance goals. This job is performed by human operators called database administrators (DBAs) who are generally well-paid, and are becoming more and more expensive with the increasing complexity and scale of modern databases. There has been considerable effort dedicated to reducing such cost (which often dominates the total ownership cost of mission-critical databases) by making database tuning more automated and transparent to users (Chaudhuri et al, 2004; Chaudhuri and Weikum, 2006). Research in this area seeks ways to automate the hardware deployment, physical database design, parameter configuration, and resource management in such systems. The goal is to achieve acceptable performance on the whole system level without (or with limited) human intervention.


2020 ◽  
Vol 9 (5) ◽  
pp. 331
Author(s):  
Dongming Guo ◽  
Erling Onstein

Geospatial information has been indispensable for many application fields, including traffic planning, urban planning, and energy management. Geospatial data are mainly stored in relational databases that have been developed over several decades, and most geographic information applications are desktop applications. With the arrival of big data, geospatial information applications are also being modified into, e.g., mobile platforms and Geospatial Web Services, which require changeable data schemas, faster query response times, and more flexible scalability than traditional spatial relational databases currently have. To respond to these new requirements, NoSQL (Not only SQL) databases are now being adopted for geospatial data storage, management, and queries. This paper reviews state-of-the-art geospatial data processing in the 10 most popular NoSQL databases. We summarize the supported geometry objects, main geometry functions, spatial indexes, query languages, and data formats of these 10 NoSQL databases. Moreover, the pros and cons of these NoSQL databases are analyzed in terms of geospatial data processing. A literature review and analysis showed that current document databases may be more suitable for massive geospatial data processing than are other NoSQL databases due to their comprehensive support for geometry objects and data formats and their performance, geospatial functions, index methods, and academic development. However, depending on the application scenarios, graph databases, key-value, and wide column databases have their own advantages.


2020 ◽  
Vol 6 (2) ◽  
pp. 187-197
Author(s):  
Nurlaila Suci Rahayu Rais ◽  
Dedeh Apriyani ◽  
Gito Gardjito

Monitoring of warehouse inventory data processing is an important thing for companies. PT Talaga mulya indah is still manual using paper media, causing problems that have an effect on existing information, namely: problems with data processing of incoming and outgoing goods. And the difference between data on the amount of stock of goods available with physical data, often occurs inputting data more than once for the same item, searching for available data, and making reports so that it impedes companies in monitoring inventory of existing stock of goods. Which aims to create a system that can provide updated information to facilitate the warehouse admin in making inventory reports, and reduce errors in input by means of integrated control. In this study, the authors used the data collection method used in this analysis using the method of observation, interviews, and literature review (literature study). For analysis using the PIECES analysis method. Furthermore, the system design used is UML (Unified Modeling Language). The results of this study are expected to produce the right data in the process of monitoring inventory data processing, also can provide the right information and make it easier to control the overall availability of goods.


2007 ◽  
Vol 105 (2) ◽  
pp. 514-522 ◽  
Author(s):  
Joy L. Hendrick ◽  
Jamie R. Switzer

As some states allow motorists to use hands-free cell phones only while driving, this study was done to examine some braking responses to see if conversing on these two types of cell phones affects quick responding. College-age drivers ( n = 25) completed reaction time trials in go/no-go situations under three conditions: control (no cell phone or conversation), and conversing on hands-free and hand-held cell phones. Their task involved moving the right foot from one pedal to another as quickly as possible in response to a visual signal in a lab setting. Significantly slower reaction times, movement times, and total response times were found for both cell phone conditions than for the control but no differences between hands-free and hand-held phone conditions. These findings provide additional support that talking on cell phones, regardless if it is hands-free or hand-held, reduces speed of information processing.


2020 ◽  
Vol 245 ◽  
pp. 05020
Author(s):  
Vardan Gyurjyan ◽  
Sebastian Mancilla

The hardware landscape used in HEP and NP is changing from homogeneous multi-core systems towards heterogeneous systems with many different computing units, each with their own characteristics. To achieve maximum performance with data processing, the main challenge is to place the right computing on the right hardware. In this paper, we discuss CLAS12 charge particle tracking workflow orchestration that allows us to utilize both CPU and GPU to improve the performance. The tracking application algorithm was decomposed into micro-services that are deployed on CPU and GPU processing units, where the best features of both are intelligently combined to achieve maximum performance. In this heterogeneous environment, CLARA aims to match the requirements of each micro-service to the strength of a CPU or a GPU architecture. A predefined execution of a micro-service on a CPU or a GPU may not be the most optimal solution due to the streaming data-quantum size and the data-quantum transfer latency between CPU and GPU. So, the CLARA workflow orchestrator is designed to dynamically assign micro-service execution to a CPU or a GPU, based on the online benchmark results analyzed for a period of real-time data-processing.


Author(s):  
Birgitta Dresp-Langley ◽  
Marie Monfouga

Pieron's and Chocholle’s seminal psychophysical work predicts that human response time to information relative to visual contrast and/or sound frequency decreases when contrast intensity or sound frequency increases. The goal of this study is to bring to the fore the ability of individuals to use visual contrast intensity and sound frequency in combination for faster perceptual decisions of relative depth (“nearer”) in planar (2D) object configurations on the basis of physical variations in luminance contrast. Computer controlled images with two abstract patterns of varying contrast intensity, one on the left and one on the right, preceded or not by a pure tone of varying frequency, were shown to healthy young humans in controlled experimental sequences. Their task (two-alternative forced-choice) was to decide as quickly as possible which of two patterns, the left or the right one, in a given image appeared to “stand out as if it were nearer” in terms of apparent (subjective) visual depth. The results show that the combinations of varying relative visual contrast with sounds of varying frequency exploited here produced an additive effect on choice response times in terms of facilitation, where a stronger visual contrast combined with a higher sound frequency produced shorter forced-choice response times. This new effect is predicted by cross-modal audio-visual probability summation.


Author(s):  
Jack Goldsmith ◽  
Tim Wu

If you had met Jon Postel in 1998, you might have been surprised to learn that you were in the presence of one of the Internet’s greatest living authorities. He had a rambling, ragged look, living in sandals and a large, unkempt beard. He lived like a modern-day Obi-Wan Kenobi, an academic hermit who favored solitary walks on the Southern California beach. When told once by a reporter that readers were interested in learning more about his personal life, he answered: “If we tell them, they won’t be interested anymore.” Yet this man was, and had been for as long as anyone could remember, the ultimate authority for assignment of the all-important Internet Protocol (IP) numbers that are the essential feature of Internet membership. Like the medallions assigned to New York City taxicabs, each globally unique number identifies a computer on the Net, determining who belongs and who doesn’t. “If the Net does have a God,” wrote the Economist in 1997, “he is probably Jon Postel.” Jon Postel was a quiet man who kept strong opinions and sometimes acted in surprising ways. The day of January 28, 1998, provided the best example. On that day Postel wrote an e-mail to the human operators of eight of the twelve “name servers” around the globe. Name servers are the critical computers that are ultimately responsible for making sure that when you type a name like google.com you reach the right address (123.23.83.0). On that day Postel asked the eight operators, all personally loyal to Postel, to recognize his computer as the “root,” or, in essence, the master computer for the whole Internet. The operators complied, pointing their servers to Postel’s computer instead of the authoritative root controlled by the United States government. The order made the operators nervous—Paul Vixie, one of the eight, quietly arranged to have someone look after his kids in case he was arrested. Postel was playing with fire. His act could have divided the Internet’s critical naming system into two gigantic networks, one headed by himself, the other headed by the United States. He engineered things so that the Internet continued to run smoothly. But had he wanted to during this critical time, he might have created chaos.


2018 ◽  
Vol 10 (3) ◽  
pp. 76-90
Author(s):  
Ye Tao ◽  
Xiaodong Wang ◽  
Xiaowei Xu

This article describes how rapidly growing data volumes require systems that have the ability to handle massive heterogeneous unstructured data sets. However, most existing mature transaction processing systems are built upon relational databases with structured data. In this article, the authors design a hybrid development framework, to offer greater scalability and flexibility of data analysis and reporting, while keeping maximum compatibility and links to the legacy platforms on which transaction business logics run. Data, service and user interfaces are implemented as a toolset stack, for developing applications with functionalities of information retrieval, data processing, analyzing and visualizing. A use case of healthcare data integration is presented as an example, where information is collected and aggregated from diverse sources. The workflow and simulation of data processing and visualization are also discussed, to validate the effectiveness of the proposed framework.


Sign in / Sign up

Export Citation Format

Share Document