Depicting Data Quality Issues in Business Intelligence Environment through a Metadata Framework

2016 ◽  
Vol 7 (2) ◽  
pp. 20-31
Author(s):  
Te-Wei Wang ◽  
Yuriy Verbitskiy ◽  
William Yeoh

Modern business intelligence systems depend highly on high quality data. The core of data quality management is to identify all possible sources of data quality problems. To achieve this goal, an extensive metadata infrastructure is the most promising solution. Through theoretical metadata model investigation, the authors identified a set of data quality dimensions by carefully examining the data quality management principles and applied those principles to current BI environment. They summarize their analysis by proposing a BI data quality framework.

Author(s):  
Te-Wei Wang ◽  
Yuriy Verbitskiy ◽  
William Yeoh

Modern business intelligence systems depend highly on high quality data. The core of data quality management is to identify all possible sources of data quality problems. To achieve this goal, an extensive metadata infrastructure is the most promising solution. Through theoretical metadata model investigation, the authors identified a set of data quality dimensions by carefully examining the data quality management principles and applied those principles to current BI environment. They summarize the analysis by proposing a BI data quality framework.


2019 ◽  
Vol 15 (2) ◽  
pp. 52-67
Author(s):  
Nori Wilantika ◽  
Wahyu Catur Wibowo

Every varsity in Indonesia is responsible for ensuring the completeness, the validity, the accuracy, and the currency of its educational data. The educational data is used for implementing higher-education quality assurance system and formulating policies related to universities and majors in Indonesia. Data quality assessment result indicates that educational data in Statistics Polytechnic did not meet completeness, validity, accuracy, and currency criteria. Data quality management maturity has been measured using Loshin’s Data Quality Maturity Model which result is in level 1 to level 2 of maturity. Only the data quality dimensions component has achieved the expected target. Thus, recommendations have been proposed based on the DAMA-DMBOK framework. The activities needed to be carried out are developing and promoting awareness of data quality; defining data quality requirements; profiling, analyzing, and evaluating data quality; define business rules for data quality, establish, and evaluate the data quality services levels, manage problems related to data quality, design and implement operational procedures for data quality management, and monitor operations and performance of data quality management procedures.


2018 ◽  
Vol 09 (01) ◽  
pp. 072-081 ◽  
Author(s):  
Lauren Houston ◽  
Yasmine Probst ◽  
Ping Yu ◽  
Allison Martin

Background Clinical trials are an important research method for improving medical knowledge and patient care. Multiple international and national guidelines stipulate the need for data quality and assurance. Many strategies and interventions are developed to reduce error in trials, including standard operating procedures, personnel training, data monitoring, and design of case report forms. However, guidelines are nonspecific in the nature and extent of necessary methods. Objective This article gathers information about current data quality tools and procedures used within Australian clinical trial sites, with the aim to develop standard data quality monitoring procedures to ensure data integrity. Methods Relevant information about data quality management methods and procedures, error levels, data monitoring, staff training, and development were collected. Staff members from 142 clinical trials listed on the National Health and Medical Research Council (NHMRC) clinical trials Web site were invited to complete a short self-reported semiquantitative anonymous online survey. Results Twenty (14%) clinical trials completed the survey. Results from the survey indicate that procedures to ensure data quality varies among clinical trial sites. Centralized monitoring (65%) was the most common procedure to ensure high-quality data. Ten (50%) trials reported having a data management plan in place and two sites utilized an error acceptance level to minimize discrepancy, set at <5% and 5 to 10%, respectively. The quantity of data variables checked (10–100%), the frequency of visits (once-a-month to annually), and types of variables (100%, critical data or critical and noncritical data audits) for data monitoring varied among respondents. The average time spent on staff training per person was 11.58 hours over a 12-month period and the type of training was diverse. Conclusion Clinical trial sites are implementing ad hoc methods pragmatically to ensure data quality. Findings highlight the necessity for further research into “standard practice” focusing on developing and implementing publicly available data quality monitoring procedures.


Author(s):  
Suranga C. H. Geekiyanage ◽  
Dan Sui ◽  
Bernt S. Aadnoy

Drilling industry operations heavily depend on digital information. Data analysis is a process of acquiring, transforming, interpreting, modelling, displaying and storing data with an aim of extracting useful information, so that the decision-making, actions executing, events detecting and incident managing of a system can be handled in an efficient and certain manner. This paper aims to provide an approach to understand, cleanse, improve and interpret the post-well or realtime data to preserve or enhance data features, like accuracy, consistency, reliability and validity. Data quality management is a process with three major phases. Phase I is an evaluation of pre-data quality to identify data issues such as missing or incomplete data, non-standard or invalid data and redundant data etc. Phase II is an implementation of different data quality managing practices such as filtering, data assimilation, and data reconciliation to improve data accuracy and discover useful information. The third and final phase is a post-data quality evaluation, which is conducted to assure data quality and enhance the system performance. In this study, a laboratory-scale drilling rig with a control system capable of drilling is utilized for data acquisition and quality improvement. Safe and efficient performance of such control system heavily relies on quality of the data obtained while drilling and its sufficient availability. Pump pressure, top-drive rotational speed, weight on bit, drill string torque and bit depth are available measurements. The data analysis is challenged by issues such as corruption of data due to noises, time delays, missing or incomplete data and external disturbances. In order to solve such issues, different data quality improvement practices are applied for the testing. These techniques help the intelligent system to achieve better decision-making and quicker fault detection. The study from the laboratory-scale drilling rig clearly demonstrates the need for a proper data quality management process and clear understanding of signal processing methods to carry out an intelligent digitalization in oil and gas industry.


Author(s):  
Tarik Chafiq ◽  
Mohammed Ouadoud ◽  
Hassane Jarar Oulidi ◽  
Ahmed Fekri

The aim of this research work is to ensure the integrity and correction of the geotechnical database which contains anomalies. These anomalies occurred mainly in the phase of inputting and/or transferring of data. The algorithm created in the framework of this paper was tested on a dataset of 70 core drillings. In fact, it is based on a multi-criteria analysis qualifying the geotechnical data integrity using the sequential approach. The implementation of this algorithm has given a relevant set of values in terms of output; which will minimalize processing time and manual verification. The application of the methodology used in this paper could be useful to define the type of foundation adapted to the nature of the subsoil, and thus, foresee the adequate budget.


Sign in / Sign up

Export Citation Format

Share Document