scholarly journals Analisa Kategori Barang dengan Penjualan Terbanyak dalam Jangka Waktu 3 Bulan Menggunakan Data Warehouse

2022 ◽  
Vol 6 (1) ◽  
pp. 65-78
Author(s):  
I Putu Agus Eka Pratama ◽  
Rey Bernard

UD. Makmur Sejahtera sebagai salah satu distributor terbesar untuk barang kebutuhan sehari-hari di Manokwari Papua, memiliki data-data transaksi penjualan untuk setiap kategori barang dan jenis barang. Data-data ini masih tersimpan secara fisik dalam bentuk nota serta belum didigitalkan untuk dapat dimanfaatkan secara maksimal untuk membantu UD. Makmur Sejahtera meningkatkan penjualan. Penelitian ini memiliki ide dasar pemanfaatan data digital transaksi penjualan untuk mengetahui kategori barang mana yang memiliki penjualan terbanyak dalam kurun waktu tiga bulan (Juli 2020 hingga September 2020) melalui proses Extraction, Transformation, Loading (ETL) berbasis Pentaho Data Integration, untuk kemudian disimpan dalam bentuk data multi dimensi, dikategorikan, dan divisualisasikan menggunakan Tableau. Hasil pengujian menunjukkan bahwa komoditas beras merupakan kategori barang dengan penjualan terbanyak pada kurun waktu tiga bulan serta implementasi Data Warehouse sangat membantu UD. Makmur Sejahtera di dalam mencapai tujuan bisnis usahanya.

2013 ◽  
Vol 321-324 ◽  
pp. 2532-2538
Author(s):  
Xiao Guo Wang ◽  
Jian Shen ◽  
Chuan Sun

Considering the difficulty of information collection and integration due to the rapid growth of information, we need an efficient tool to do these jobs. A proposal is be put forward to build a data integration system to collect the source data and preprocess the heterogeneous data and then convert/extract data to the data warehouse. Through experiment and analysis, this paper designed an information process flow and implemented the data integration system, based on B/S framework with the database technology, to deal with the college related information.


2014 ◽  
Vol 556-562 ◽  
pp. 5783-5786
Author(s):  
Wan Chun Zhu

In this paper, on the basis of the data warehouse integration mechanisms, use Object-oriented analysis methods, take the ITIS’s functions and data needs as the starting point, promote and optimize the ETL method and form a Tourism-oriented ITIS data integration model. The “wisdom tourism” is explained elaborately. Then it shows what data integration is and how to use integration.


2015 ◽  
Vol 78 (1) ◽  
Author(s):  
Mimi Safinaz Jamaluddin ◽  
Nurulhuda Firdaus Mohd Azmi

Data integration is important in consolidating all the data in the organization or outside the organization to provide a unified view of the organization's information. Extraction Transformation Load (ETL) solution is the back-end process of data integration which involves collecting data from various data sources, preparing and transforming the data according to business requirements and loading them into a Data Warehouse (DW). This paper explains the integration of the rubber import and export data between Malaysian Rubber Board (MRB) and Royal Malaysian Customs Department (Customs) using the ETL solution. Microsoft SQL Server Integration Services (SSIS) and Microsoft SQL Server Agent Jobs have been used as the ETL tool and ETL scheduling. 


2010 ◽  
Vol 7 (3) ◽  
Author(s):  
Giorgio Ghisalberti ◽  
Marco Masseroli ◽  
Luca Tettamanti

SummaryNumerous biomolecular data are available, but they are scattered in many databases and only some of them are curated by experts. Most available data are computationally derived and include errors and inconsistencies. Effective use of available data in order to derive new knowledge hence requires data integration and quality improvement. Many approaches for data integration have been proposed. Data warehousing seams to be the most adequate when comprehensive analysis of integrated data is required. This makes it the most suitable also to implement comprehensive quality controls on integrated data. We previously developed GFINDer (http://www.bioinformatics.polimi.it/GFINDer/), a web system that supports scientists in effectively using available information. It allows comprehensive statistical analysis and mining of functional and phenotypic annotations of gene lists, such as those identified by high-throughput biomolecular experiments. GFINDer backend is composed of a multi-organism genomic and proteomic data warehouse (GPDW). Within the GPDW, several controlled terminologies and ontologies, which describe gene and gene product related biomolecular processes, functions and phenotypes, are imported and integrated, together with their associations with genes and proteins of several organisms. In order to ease maintaining updated the GPDW and to ensure the best possible quality of data integrated in subsequent updating of the data warehouse, we developed several automatic procedures. Within them, we implemented numerous data quality control techniques to test the integrated data for a variety of possible errors and inconsistencies. Among other features, the implemented controls check data structure and completeness, ontological data consistency, ID format and evolution, unexpected data quantification values, and consistency of data from single and multiple sources. We use the implemented controls to analyze the quality of data available from several different biological databases and integrated in the GFINDer data warehouse. By doing so, we identified in these data a variety of different types of errors and inconsistencies; this enables us to ensure good quality of the data in the GFINDer data warehouse. We reported all identified data errors and inconsistencies to the curators of the original databases from where the data were retrieved, who mainly corrected them in subsequent updating of the original database. This contributed to improve the quality of the data available, in the original databases, to the whole scientific community.


Author(s):  
Raphael W. Majeed ◽  
Mark R. Stöhr ◽  
Andreas Günther

Data integration is a necessary and important step to perform translational research and improve the sample size beyond single data collections. For health information, the most recent established communication standards is HL7 FHIR. To bridge the concepts of “minimal invasive” data integration and open standards, we propose a generic ETL framework to process arbitrary patient related data collections into HL7 FHIR – which in turn can then be used for loading into target data warehouses. The proposed algorithm is able to read any relational delimited text exports and produce a standard HL7 FHIR bundle collection. We evaluated an implementation of the algorithm using different lung research registries and used the resulting FHIR resources to fill our i2b2 based data warehouse as well an OMOP common data model repository.


Author(s):  
Tom Breur

Business Intelligence (BI) projects that involve substantial data integration have often proven failure-prone and difficult to plan. Data quality issues trigger rework, which makes it difficult to accurately schedule deliverables. Two things can bring improvement. Firstly, one should deliver information products in the smallest possible chunks, but without adding prohibitive overhead for breaking up the work in tiny increments. This will increase the frequency and improve timeliness of feedback on suitability of information products and hence make planning and progress more predictable. Secondly, BI teams need to provide better stewardship when they facilitate discussions between departments whose data cannot easily be integrated. Many so-called data quality errors do not stem from inaccurate source data, but rather from incorrect interpretation of data. This is mostly caused by different interpretation of essentially the same underlying source system facts across departments with misaligned performance objectives. Such problems require prudent stakeholder management and informed negotiations to resolve such differences. In this chapter, the authors suggest an innovation to data warehouse architecture to help accomplish these objectives.


Sign in / Sign up

Export Citation Format

Share Document