scholarly journals Value of big data to finance: observations on an internet credit Service Company in China

2015 ◽  
Vol 1 (1) ◽  
Author(s):  
Shaofeng Zhang ◽  
Wei Xiong ◽  
Wancheng Ni ◽  
Xin Li

Abstract Background his paper presents a case study on 100Credit, an Internet credit service provider in China. 100Credit began as an IT company specializing in e-commerce recommendation before getting into the credit rating business. The company makes use of Big Data on multiple aspects of individuals’ online activities to infer their potential credit risk. Methods Based on 100Credit’s business practices, this paper summarizes four aspects related to the value of Big Data in Internet credit services. Results 1) value from large data volume that provides access to more borrowers; 2) value from prediction correctness in reducing lenders’ operational cost; 3) value from the variety of services catering to different needs of lenders; and 4) value from information protection to sustain credit service businesses. Conclusion The paper also discusses the opportunities and challenges of Big Data-based credit risk analysis, which needs to be improved in future research and practice.

2020 ◽  
pp. 275-348
Author(s):  
Terence M. Yhip ◽  
Bijan M. D. Alagheband

2021 ◽  
Vol 8 (2) ◽  
pp. 205395172110481
Author(s):  
Remy Stewart

Consumer-based datasets are the products of data brokerage firms that agglomerate millions of personal records on the adult US population. This big data commodity is purchased by both companies and individual clients for purposes such as marketing, risk prevention, and identity searches. The sheer magnitude and population coverage of available consumer-based datasets and the opacity of the business practices that create these datasets pose emergent ethical challenges within the computational social sciences that have begun to incorporate consumer-based datasets into empirical research. To directly engage with the core ethical debates around the use of consumer-based datasets within social science research, I first consider two case study applications of consumer-based dataset-based scholarship. I then focus on three primary ethical dilemmas within consumer-based datasets regarding human subject research, participant privacy, and informed consent in conversation with the principles of the seminal Belmont Report.


Big Data ◽  
2016 ◽  
pp. 2249-2274
Author(s):  
Chinh Nguyen ◽  
Rosemary Stockdale ◽  
Helana Scheepers ◽  
Jason Sargent

The rapid development of technology and interactive nature of Government 2.0 (Gov 2.0) is generating large data sets for Government, resulting in a struggle to control, manage, and extract the right information. Therefore, research into these large data sets (termed Big Data) has become necessary. Governments are now spending significant finances on storing and processing vast amounts of information because of the huge proliferation and complexity of Big Data and a lack of effective records management. On the other hand, there is a method called Electronic Records Management (ERM), for controlling and governing the important data of an organisation. This paper investigates the challenges identified from reviewing the literature for Gov 2.0, Big Data, and ERM in order to develop a better understanding of the application of ERM to Big Data to extract useable information in the context of Gov 2.0. The paper suggests that a key building block in providing useable information to stakeholders could potentially be ERM with its well established governance policies. A framework is constructed to illustrate how ERM can play a role in the context of Gov 2.0. Future research is necessary to address the specific constraints and expectations placed on governments in terms of data retention and use.


Author(s):  
Arun Thotapalli Sundararaman

Study of data quality for data mining application has always been a complex topic; in the recent years, this topic has gained further complexity with the advent of big data as the source for data mining and business intelligence (BI) applications. In a big data environment, data is consumed in various states and various forms serving as input for data mining, and this is the main source of added complexity. These new complexities and challenges arise from the underlying dimensions of big data (volume, variety, velocity, and value) together with the ability to consume data at various stages of transition from raw data to standardized datasets. These have created a need for expanding the traditional data quality (DQ) factors into BDQ (big data quality) factors besides the need for new BDQ assessment and measurement frameworks for data mining and BI applications. However, very limited advancement has been made in research and industry in the topic of BDQ and their relevance and criticality for data mining and BI applications. Data quality in data mining refers to the quality of the patterns or results of the models built using mining algorithms. DQ for data mining in business intelligence applications should be aligned with the objectives of the BI application. Objective measures, training/modeling approaches, and subjective measures are three major approaches that exist to measure DQ for data mining. However, there is no agreement yet on definitions or measurements or interpretations of DQ for data mining. Defining the factors of DQ for data mining and their measurement for a BI system has been one of the major challenges for researchers as well as practitioners. This chapter provides an overview of existing research in the area of BDQ definitions and measurement for data mining for BI, analyzes the gaps therein, and provides a direction for future research and practice in this area.


2019 ◽  
pp. 089443931988845 ◽  
Author(s):  
Alexander Christ ◽  
Marcus Penthin ◽  
Stephan Kröner

Systematic reviews are the method of choice to synthesize research evidence. To identify main topics (so-called hot spots) relevant to large corpora of original publications in need of a synthesis, one must address the “three Vs” of big data (volume, velocity, and variety), especially in loosely defined or fragmented disciplines. For this purpose, text mining and predictive modeling are very helpful. Thus, we applied these methods to a compilation of documents related to digitalization in aesthetic, arts, and cultural education, as a prototypical, loosely defined, fragmented discipline, and particularly to quantitative research within it (QRD-ACE). By broadly querying the abstract and citation database Scopus with terms indicative of QRD-ACE, we identified a corpus of N = 55,553 publications for the years 2013–2017. As the result of an iterative approach of text mining, priority screening, and predictive modeling, we identified n = 8,304 potentially relevant publications of which n = 1,666 were included after priority screening. Analysis of the subject distribution of the included publications revealed video games as a first hot spot of QRD-ACE. Topic modeling resulted in aesthetics and cultural activities on social media as a second hot spot, related to 4 of k = 8 identified topics. This way, we were able to identify current hot spots of QRD-ACE by screening less than 15% of the corpus. We discuss implications for harnessing text mining, predictive modeling, and priority screening in future research syntheses and avenues for future original research on QRD-ACE.


2016 ◽  
Vol 20 (1) ◽  
pp. 12-28 ◽  
Author(s):  
Son K. Lam ◽  
Stefan Sleep ◽  
Thorsten Hennig-Thurau ◽  
Shrihari Sridhar ◽  
Alok R. Saboo

The advent of new forms of data, modern technology, and advanced data analytics offer service providers both opportunities and risks. This article builds on the phenomenon of big data and offers an integrative conceptual framework that captures not only the benefits but also the costs of big data for managing the frontline employee (FLE)-customer interaction. Along the positive path, the framework explains how the “3Vs” of big data (volume, velocity, and variety) have the potential to improve service quality and reduce service costs by influencing big data value and organizational change at the firm and FLE levels. However, the 3Vs of big data also increase big data veracity, which casts doubt about the value of big data. The authors further propose that because of heterogeneity in big data absorptive capacities at the firm level, the costs of adopting big data in FLE management may outweigh the benefits. Finally, while FLEs can benefit from big data, extracting knowledge from such data does not discount knowledge derived from FLEs’ small data. Rather, combining and integrating the firm’s big data with FLEs’ small data are crucial to absorbing and applying big data knowledge. An agenda for future research concludes.


Author(s):  
A. Sheik Abdullah ◽  
R. Suganya ◽  
S. Selvakumar ◽  
S. Rajaram

Classification is considered to be the one of the data analysis technique which can be used over many applications. Classification model predicts categorical continuous class labels. Clustering mainly deals with grouping of variables based upon similar characteristics. Classification models are experienced by comparing the predicted values to that of the known target values in a set of test data. Data classification has many applications in business modeling, marketing analysis, credit risk analysis; biomedical engineering and drug retort modeling. The extension of data analysis and classification makes the insight into big data with an exploration to processing and managing large data sets. This chapter deals with various techniques, methodologies that correspond to the classification problem in data analysis process and its methodological impacts to big data.


2019 ◽  
Vol 26 ◽  
pp. 03002
Author(s):  
Tilei Gao ◽  
Ming Yang ◽  
Rong Jiang ◽  
Yu Li ◽  
Yao Yao

The emergence of big data has brought a great impact on traditional computing mode, the distributed computing framework represented by MapReduce has become an important solution to this problem. Based on the big data, this paper deeply studies the principle and framework of MapReduce programming. On the basis of mastering the principle and framework of MapReduce programming, the time consumption of distributed computing framework MapReduce and traditional computing model is compared with concrete programming experiments. The experiment shows that MapReduce has great advantages in large data volume.


Author(s):  
E. B. Nizamieva

Purpose: The aim of this work is to show how smart cities can drive the reorganization and efficiency of existing cities.Design/methodology/approach: The paper describes modern achievements in the field of a smart city, the latest achievements of cities and technological solutions they introduce. The paper analyzes when and why this concept appears, development stages and prospects of this concept. The world problems of the urbanization process in new territories and ways to solve them.Research findings: The paper considers relevant reports and studies highlighting the problems and solutions of urbanization and the ecological situation in cities, the negative impact on the environment.Practical implications: One of the ways to solve such problems is the implementation of a set of solutions included in the smart city concept. How modern technological solutions and large data volume assist in the communal and economic resource management, overcome environmental challenges of today and make the city more accessible to its residents. How historical cities can actively integrate and improve urban environment with minimal intervention.Originality/value: Attempts are made to answer whether cities need to become smart, what the consequences may be. As a consequence of emerging issues, many problem must be discussed in future research.


2020 ◽  
Vol 10 (2) ◽  
pp. 7-10
Author(s):  
Deepti Pandey

This article provides insight into an emerging research discipline called Psychoinformatics.In the context of Psychoinformatics, we emphasize the co-operation between the disciplines of Psychology and Information Science which handles large data sets is derivative from severely used devices like smartphones or any online social networking in order to highlight  sychological qualities including both personality and mood. New challenges await psychologists considering the result “Big Data” sets because classic psychological methods will only in part be able to analyze this data derived from ubiquitous mobile devices as well as other everyday technologies. Consequently, psychologist must enrich their scientific methods through the inclusion of methods from informatics. Furthermore, we also emphasize on data which is derived from Psychoinformatics to combine in a such a way to give meaningful way with data from human neuroscience. We close the article with some observations of areas for future research and problems that require consideration within this new discipline.


Sign in / Sign up

Export Citation Format

Share Document