A Big Data Query Optimization Framework for Telecom Customer Churn Analysis

2021 ◽  
pp. 475-484
Author(s):  
Aarti Chugh ◽  
Vivek Kumar Sharma ◽  
Manjot Kaur Bhatia ◽  
Charu Jain
2016 ◽  
Vol 9 (12) ◽  
pp. 1005-1016 ◽  
Author(s):  
Hai Liu ◽  
Dongqing Xiao ◽  
Pankaj Didwania ◽  
Mohamed Y. Eltabakh

2021 ◽  
Author(s):  
Anuja S. ◽  
Malathy C.

Abstract In today's world, most of the private and public sector organizations deal with massive amounts of raw data, which includes information and knowledge in their secret layer. In addition, the format, scale, variety, and velocity of generated data make it more difficult to use the algorithms in an efficient manner. This complexity necessitates the use of sophisticated methods, strategies, and algorithms to solve the challenges of managing raw data. Big data query optimization (BDQO) requires businesses to define, diagnose, forecast, prescribe, and cognize hidden growth opportunities and guiding them toward achieving market value. BDQO uses advanced analytical methods to extract information from an increasingly growing volume of data, resulting in a reduction in the difficulty of the decision-making process. Hadoop, Apache Hive, No SQL, Map Reduce, and HPCC are the technologies used in big data applications to manage large data. It is less costly to consume data for query processing because big data provides scalability. However, small businesses will never be able to query large databases. Joining tables with millions of tuples could take hours. Parallelism, which solves the problem by using more processors, may be a potential solution. Unfortunately, small businesses cannot afford to operate on a shoestring budget. There are many techniques to tackle the problem. The technologies used in the big data query optimization process are discussed in depth in this paper.


2019 ◽  
Vol 16 (8) ◽  
pp. 3428-3431
Author(s):  
Kamya Eria ◽  
Booma Poolan Marikannan

The telecom industry is saturated with many service providers competing for highly rational customers. The current big data and highly technological era calls for real-time churn analysis and decision making which has also been highlighted in previous studies. However, telecom data is highly dimensional in nature thus when this is coupled with this big data era increases the computational and processing costs. Therefore, this complexity and dimensionality of telecom data coupled with the current need for near or real-time churn analysis demands feature selection-based models that efficiently consider the most relevant variables in explaining customer churn behaviors. This study proposes a feature extraction-based churn prediction model that concentrates on the most relevant features with significant discriminatory power for churn. The data has been reduced on the basis of missing values and irrelevant variables. Irrelevant variables were first identified by use of Random Forest and Logistic Regression models. The findings of the study provide churn analysts with insights about the prediction errors to consider and minimize in their future churn analyses. It also contributes to reducing computational costs incurred by churn analysts working with big data in their churn prediction and analysis.


2018 ◽  
Vol 27 (6) ◽  
pp. 873-898 ◽  
Author(s):  
Yuchen Liu ◽  
Hai Liu ◽  
Dongqing Xiao ◽  
Mohamed Y. Eltabakh

Author(s):  
Pankaj Dadheech ◽  
Dinesh Goyal ◽  
Sumit Srivastava ◽  
Ankit Kumar

Spatial queries frequently used in Hadoop for significant data process. However, vast and massive size of spatial information makes it difficult to process the spatial inquiries proficiently, so they utilized the Hadoop system for process Big Data. We have used Boolean Queries & Geometry Boolean Spatial Data for Query Optimization using Hadoop System. In this paper, we show a lightweight and adaptable spatial data index for big data which will process in Hadoop frameworks. Results demonstrate the proficiency and adequacy of our spatial ordering system for various spatial inquiries.


2013 ◽  
Vol 756-759 ◽  
pp. 916-921
Author(s):  
Ye Liang

The amount of data in our industry and the world is exploding. Data is being collected and stored at unprecedented rates. The challenge is not only to store and manage the vast volume of data, which is also called big data, but also to analyze and query from it. In order to put forward the universal method to response mobile big data query, queries are separated and grouped according to kinds of query for massive mobile objects in the space. The indexing method for grouping the mobile objects with Grid (GG TPR-tree) has great efficiency to manage a massive capacity of mobile objects within a limited area, but it only could meet a part of requirements for mobile big data query if the GG TPR-tree was used solely. This thesis offers solutions to simple immediate query, simple continuous query, active window query, and continuous window query, dynamic condition query and other query requests by employing DTDI index structure. The experiments prove that with the support of DTDI index structure, query of massive mobile objects has higher precision and better query performance.


Author(s):  
Jintao Gao ◽  
Zhanhuai Li ◽  
Wenjie Liu

Cardinality estimation is an important component of query optimization. Its accuracy and efficiency directly decide effect of query optimization. Traditional cardinality estimation strategy is based on original table or sample to collect statistics, then inferring cardinality by collected statistics. It will be low-efficiency when handling big data; Statistics exist update latency and are gotten by inferring, which can not guarantee correctness; Some strategies can get the actual cardinality by executing some subqueries, but they do not keep the result, leading to low efficiency of fetching statistics. Against these problems, this paper proposes a novel cardinality estimation strategy, called cardinality estimation based on query result(CEQR). For keeping correctness of cardinality, CEQR directly gets statistics from query results, which is not related with data size; we build a cardinality table to store the statistics of basic tables and middle results under specific predicates. Cardinality table can provide cardinality services for subsequent queries, and we build a suit of rules to maintain cardinality table; To improve the efficiency of fetching statistics, we introduce the source aware strategy, which hashes cardinality item to appropriate cache. This paper gives the adaptability and deviation analytic of CEQR, and proves that CEQR is more efficient than traditional cardinality estimation strategy by experiments.


Sign in / Sign up

Export Citation Format

Share Document