scholarly journals The Model Design of Medical Data Life Cycle Based on Big Data Platform

2021 ◽  
Vol 1865 (4) ◽  
pp. 042088
Author(s):  
Ziqing Li ◽  
Shenglei Pei ◽  
Guiliang Feng
2020 ◽  
Vol 173 ◽  
pp. 364-371
Author(s):  
Kumar Rahul ◽  
Rohitash Kumar Banyal

2018 ◽  
Vol 26 (1) ◽  
pp. 153-170 ◽  
Author(s):  
Emily M. Coyne ◽  
Joshua G. Coyne ◽  
Kenton B. Walker

Purpose Big Data has become increasingly important to multiple facets of the accounting profession, but accountants have little understanding of the steps necessary to convert Big Data into useful information. This limited understanding creates a gap between what accountants can do and what accountants should do to assist in Big Data information governance. The study aims to bridge this gap in two ways. Design/methodology/approach First, the study introduces a model of the Big Data life cycle to explain the process of converting Big Data into information. Knowledge of this life cycle is a first step toward enabling accountants to engage in Big Data information governance. Second, it highlights informational and control risks inherent to this life cycle, and identifies information governance activities and agents that can minimize these risks. Findings Because accountants have a strong ability to identify the informational and control needs of internal and external decision-makers, they should play a significant role in Big Data information governance. Originality/value This model of the Big Data life cycle and information governance provides a first attempt to formalize knowledge that accountants need in a new field of the accounting profession.


2021 ◽  
Author(s):  
Victoria Tokareva ◽  
Igor Bychkov ◽  
Andrey Demichev ◽  
Julia Dubenskaya ◽  
Oleg Fedorov ◽  
...  

2018 ◽  
Vol 7 (1) ◽  
pp. 33-39
Author(s):  
Kanika . ◽  
Alka . ◽  
R. A. Khan

Big data is a huge amount of data created by individuals related to their medical, internet activity, social networking sites, energy usage communication patterns etc. From these sources, data is being collected and processed by various survey organizations, national statistical agencies, medical centres, and other companies etc. There are many security challenges which occur during data transactions, such as un-authentication, phishing, Vishing, data mining based attacks, etc. From a security point of view the biggest challenge for big data is the protection of user’s privacy. Yazan et.al, have presented big data lifecycle threat model. This paper does a critical review of the work. An Improved Security Threat Model for Big Data Life Cycle has been proposed as a main contribution of the paper. A new phase i.e. data creation phase has been added to the life cycle and it is claimed that the phase is very important one with respect to security and privacy. To justify the claim theoretical and statistical evidences have been provided.


2020 ◽  
Vol 29 (04) ◽  
pp. 2030001
Author(s):  
Martin Macak ◽  
Mouzhi Ge ◽  
Barbora Buhnova

Nowadays, a variety of Big Data architectures are emerging to organize the Big Data life cycle. While some of these architectures are proposed for general usage, many of them are proposed in a specific application domain such as smart cities, transportation, healthcare, and agriculture. There is, however, a lack of understanding of how and why Big Data architectures vary in different domains and how the Big Data architecture strategy in one domain may possibly advance other domains. Therefore, this paper surveys and compares the Big Data architectures in different application domains. It also chooses a representative architecture of each researched application domain to indicate which Big Data architecture from a given domain the researchers and practitioners may possibly start from. Next, a pairwise cross-domain comparison among the Big Data architectures is presented to outline the similarities and differences between the domain-specific architectures. Finally, the paper provides a set of practical guidelines for Big Data researchers and practitioners to build and improve Big Data architectures based on the knowledge gathered in this study.


2016 ◽  
Vol 10 (2) ◽  
pp. 176-192 ◽  
Author(s):  
Line Pouchard

As science becomes more data-intensive and collaborative, researchers increasingly use larger and more complex data to answer research questions. The capacity of storage infrastructure, the increased sophistication and deployment of sensors, the ubiquitous availability of computer clusters, the development of new analysis techniques, and larger collaborations allow researchers to address grand societal challenges in a way that is unprecedented. In parallel, research data repositories have been built to host research data in response to the requirements of sponsors that research data be publicly available. Libraries are re-inventing themselves to respond to a growing demand to manage, store, curate and preserve the data produced in the course of publicly funded research. As librarians and data managers are developing the tools and knowledge they need to meet these new expectations, they inevitably encounter conversations around Big Data. This paper explores definitions of Big Data that have coalesced in the last decade around four commonly mentioned characteristics: volume, variety, velocity, and veracity. We highlight the issues associated with each characteristic, particularly their impact on data management and curation. We use the methodological framework of the data life cycle model, assessing two models developed in the context of Big Data projects and find them lacking. We propose a Big Data life cycle model that includes activities focused on Big Data and more closely integrates curation with the research life cycle. These activities include planning, acquiring, preparing, analyzing, preserving, and discovering, with describing the data and assuring quality being an integral part of each activity. We discuss the relationship between institutional data curation repositories and new long-term data resources associated with high performance computing centers, and reproducibility in computational science. We apply this model by mapping the four characteristics of Big Data outlined above to each of the activities in the model. This mapping produces a set of questions that practitioners should be asking in a Big Data project


Sign in / Sign up

Export Citation Format

Share Document