Effects of app pricing structures on product evaluations

2020 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Marcus Renatus Johannes Wolkenfelt ◽  
Frederik Bungaran Ishak Situmeang

Purpose The purpose of this paper is to contribute to the marketing literature and practice by examining the effect of product pricing on consumer behaviours with regard to the assertiveness and the sentiments expressed in their product reviews. In addition, the paper uses new data collection and machine learning tools that can also be extended for other research of online consumer reviewing behaviours. Design/methodology/approach Using web crawling techniques, a large data set was extracted from the Google Play Store. Following this, the authors created machine learning algorithms to identify topics from product reviews and to quantify assertiveness and sentiments from the review texts. Findings The results indicate that product pricing models affect consumer review sentiment, assertiveness and topics. Removing upfront payment obligations positively impacts the overall and pricing specific consumer sentiment and reduces assertiveness. Research limitations/implications The results reveal new effects of pricing models on the nature of consumer reviews of products and form a basis for future research. The study was conducted in the gaming category of the Google Play Store and the generalisability of the findings for other app segments or marketplaces should be further tested. Originality/value The findings can help companies that create digital products in choosing a pricing strategy for their apps. The paper is the first to investigate how pricing modes affect the nature of online reviews written by consumers.

2022 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Krishnadas Nanath ◽  
Supriya Kaitheri ◽  
Sonia Malik ◽  
Shahid Mustafa

Purpose The purpose of this paper is to examine the factors that significantly affect the prediction of fake news from the virality theory perspective. The paper looks at a mix of emotion-driven content, sentimental resonance, topic modeling and linguistic features of news articles to predict the probability of fake news. Design/methodology/approach A data set of over 12,000 articles was chosen to develop a model for fake news detection. Machine learning algorithms and natural language processing techniques were used to handle big data with efficiency. Lexicon-based emotion analysis provided eight kinds of emotions used in the article text. The cluster of topics was extracted using topic modeling (five topics), while sentiment analysis provided the resonance between the title and the text. Linguistic features were added to the coding outcomes to develop a logistic regression predictive model for testing the significant variables. Other machine learning algorithms were also executed and compared. Findings The results revealed that positive emotions in a text lower the probability of news being fake. It was also found that sensational content like illegal activities and crime-related content were associated with fake news. The news title and the text exhibiting similar sentiments were found to be having lower chances of being fake. News titles with more words and content with fewer words were found to impact fake news detection significantly. Practical implications Several systems and social media platforms today are trying to implement fake news detection methods to filter the content. This research provides exciting parameters from a viral theory perspective that could help develop automated fake news detectors. Originality/value While several studies have explored fake news detection, this study uses a new perspective on viral theory. It also introduces new parameters like sentimental resonance that could help predict fake news. This study deals with an extensive data set and uses advanced natural language processing to automate the coding techniques in developing the prediction model.


2019 ◽  
Vol 119 (1) ◽  
pp. 129-147 ◽  
Author(s):  
Pengfei Zhao ◽  
Ji Wu ◽  
Zhongsheng Hua ◽  
Shijian Fang

PurposeThe purpose of this paper is to identify electronic word-of-mouth (eWOM) customers from customer reviews. Thus, firms can precisely leverage eWOM customers to increase their product sales.Design/methodology/approachThis research proposed a framework to analyze the content of consumer-generated product reviews. Specific algorithms were used to identify potential eWOM reviewers, and then an evaluation method was used to validate the relationship between product sales and the eWOM reviewers identified by the authors’ proposed method.FindingsThe results corroborate that online product reviews that are made by the eWOM customers identified by the authors’ proposed method are more related to product sales than customer reviews that are made by non-eWOM customers and that the predictive power of the reviews generated by eWOM customers are significantly higher than the reviews generated by non-eWOM customers.Research limitations/implicationsThe proposed method is useful in the data set, which is based on one type of products. However, for other products, the validity must be tested. Previous eWOM customers may have no significant influence on product sales in the future. Therefore, the proposed method should be tested in the new market environment.Practical implicationsBy combining the method with the previous customer segmentation method, a new framework of customer segmentation is proposed to help firms understand customers’ value specifically.Originality/valueThis study is the first to identify eWOM customers from online reviews and to evaluate the relationship between reviewers and product sales.


2020 ◽  
Author(s):  
Renato Cordeiro de Amorim

In a real-world data set there is always the possibility, rather high in our opinion, that different features may have different degrees of relevance. Most machine learning algorithms deal with this fact by either selecting or deselecting features in the data preprocessing phase. However, we maintain that even among relevant features there may be different degrees of relevance, and this should be taken into account during the clustering process. With over 50 years of history, K-Means is arguably the most popular partitional clustering algorithm there is. The first K-Means based clustering algorithm to compute feature weights was designed just over 30 years ago. Various such algorithms have been designed since but there has not been, to our knowledge, a survey integrating empirical evidence of cluster recovery ability, common flaws, and possible directions for future research. This paper elaborates on the concept of feature weighting and addresses these issues by critically analysing some of the most popular, or innovative, feature weighting mechanisms based in K-Means


2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Alekh Gour ◽  
Shikha Aggarwal ◽  
Mehmet Erdem

Purpose The dynamic yet volatile nature of tourism and travel industry in a competitive environment calls for enhanced marketing intelligence and analytics, especially for those entities with limited marketing budgets. The past decade has witnessed an increased use of user-generated content (UGC) analysis as a marketing tool to make better informed decisions. Likewise, textual data analysis of UGC has gained much attention among tourism and hospitality scholars. Nonetheless, most of the scholarly works have focused on the singular application of an existing method or technique rather than using a multi-method approach. The purpose of this study is to propose a novel Web analytics methodology to examine online reviews posted by tourists in real time and assist decision-makers tasked with marketing strategy and intelligence. Design/methodology/approach For illustration, the case of tourism campaign in India was undertaken. A total of 305,298 reviews were collected, and after filtering, 276,154 reviews were qualified for analysis using a string of models. Descriptive charts, sentiment analysis, clustering, topic modeling and machine learning algorithms for real-time classification were applied. Findings Using big data from TripAdvisor, a total of 145 tourist destinations were clustered based on tourists’ perceptions. Further exploration of each cluster through topic modeling was conducted, which revealed interesting insights into satisfiers and dissatisfiers of different clusters of destinations. The results supported the use of the proposed multi-method Web-analytics approach. Practical implications The proposed machine learning model demonstrated that it could provide real-time information on the sentiments in each incoming review about a destination. This information might be useful for taking timely action for improvisation or controlling a service situation. Originality/value In terms of Web-analytics and UGC, a comprehensive analytical model to perform an end-to-end understanding of tourist behavior patterns and offer the potential for real-time interpretation is rarely proposed. The current study not only proposes such a model but also offers empirical evidence for a successful application. It contributes to the literature by providing scholars interested in textual analytics a step-by-step guide to implement a multi-method approach.


2015 ◽  
Vol 82 (4) ◽  
pp. 992-1003 ◽  
Author(s):  
Eric D. Becraft ◽  
Jeremy A. Dodsworth ◽  
Senthil K. Murugapiran ◽  
J. Ingemar Ohlsson ◽  
Brandon R. Briggs ◽  
...  

ABSTRACTThe vast majority of microbial life remains uncatalogued due to the inability to cultivate these organisms in the laboratory. This “microbial dark matter” represents a substantial portion of the tree of life and of the populations that contribute to chemical cycling in many ecosystems. In this work, we leveraged an existing single-cell genomic data set representing the candidate bacterial phylum “Calescamantes” (EM19) to calibrate machine learning algorithms and define metagenomic bins directly from pyrosequencing reads derived from Great Boiling Spring in the U.S. Great Basin. Compared to other assembly-based methods, taxonomic binning with a read-based machine learning approach yielded final assemblies with the highest predicted genome completeness of any method tested. Read-first binning subsequently was used to extractCalescamantesbins from all metagenomes with abundantCalescamantespopulations, including metagenomes from Octopus Spring and Bison Pool in Yellowstone National Park and Gongxiaoshe Spring in Yunnan Province, China. Metabolic reconstruction suggests thatCalescamantesare heterotrophic, facultative anaerobes, which can utilize oxidized nitrogen sources as terminal electron acceptors for respiration in the absence of oxygen and use proteins as their primary carbon source. Despite their phylogenetic divergence, the geographically separateCalescamantespopulations were highly similar in their predicted metabolic capabilities and core gene content, respiring O2, or oxidized nitrogen species for energy conservation in distant but chemically similar hot springs.


2017 ◽  
Vol 117 (8) ◽  
pp. 1687-1706
Author(s):  
Daeseon Choi ◽  
Younho Lee ◽  
Seokhyun Kim ◽  
Pilsung Kang

Purpose As the number of users on social network services (SNSs) continues to increase at a remarkable rate, privacy and security issues are consistently arising. Although users may not want to disclose their private attributes, these can be inferred from their public behavior on social media. In order to investigate the severity of the leakage of private information in this manner, the purpose of this paper is to present a method to infer undisclosed personal attributes of users based only on the data available on their public profiles on Facebook. Design/methodology/approach Facebook profile data consisting of 32 attributes were collected for 111,123 Korean users. Inferences were made for four private attributes (gender, age, marital status, and relationship status) based on five machine learning-based classification algorithms and three regression algorithms. Findings Experimental results showed that users’ gender can be inferred very accurately, whereas marital status and relationship status can be predicted more accurately with the authors’ algorithms than with a random model. Moreover, the average difference between the actual and predicted ages of users was only 0.5 years. The results show that some private attributes can be easily inferred from only a few pieces of user profile information, which can jeopardize personal information and may increase the risk to dignity. Research limitations/implications In this paper, the authors’ only utilized each user’s own profile data, especially text information. Since users in SNSs are directly or indirectly connected, inference performance can be improved if the profile data of the friends of a given user are additionally considered. Moreover, utilizing non-text profile information, such as profile images, can help increase inference accuracy. The authors’ can also provide a more generalized inference performance if a larger data set of Facebook users is available. Practical implications A private attribute leakage alarm system based on the inference model would be helpful for users not desirous of the disclosure of their private attributes on SNSs. SNS service providers can measure and monitor the risk of privacy leakage in their system to protect their users and optimize the target marketing based on the inferred information if users agree to use it. Originality/value This paper investigates whether private attributes of SNS users can be inferred with a few pieces of publicly available information although users are not willing to disclose them. The experimental results showed that gender, age, marital status, and relationship status, can be inferred by machine-learning algorithms. Based on these results, an early warning system was designed to help both service providers and users to protect the users’ privacy.


Author(s):  
Mervin Joe Thomas ◽  
Mithun M. Sanjeev ◽  
A.P. Sudheer ◽  
Joy M.L.

Purpose This paper aims to use different machine learning (ML) algorithms for the prediction of inverse kinematic solutions in parallel manipulators (PMs) to overcome the computational difficulties and approximations involved with the analytical methods. The results obtained from the ML algorithms and the Denavit–Hartenberg (DH) approach are compared with the experimental results to evaluate their performances. The study is performed on a novel 6-degree of freedom (DoF) PM that offers precise motions with a large workspace for the end effector. Design/methodology/approach The kinematic model for the proposed 3-PPSS PM is obtained using the modified DH approach and its inverse kinematic solutions are determined using the Levenberg–Marquardt algorithm. Various prediction algorithms such as the multiple linear regression, multi-variate polynomial regression, support vector, decision tree, random forest regression and multi-layer perceptron networks are applied to predict the inverse kinematic solutions for the manipulator. The data set required to train the network is generated experimentally by recording the poses of the end effector for different instantaneous positions of the slider using the concept of ArUco markers. Findings This paper fully demonstrates the possibility to use artificial intelligence for the prediction of inverse kinematic solutions especially for complex geometries. Originality/value As the analytical models derived from the geometrical method, Screw theory or numerical techniques involve approximations and needs more computational power, it is not advisable for real-time control of the manipulator. In addition, the data set obtained from the derived inverse kinematic equations to train the network may lead to inaccuracies in the predicted results. This error may generate significant deviations in the end-effector position from the desired position. The present work attempts to resolve this issue by proposing a camera-based approach that uses ArUco library and ML algorithms to create the data set experimentally and predict the inverse kinematic solutions accurately.


2019 ◽  
Vol 15 (5) ◽  
pp. 489-509 ◽  
Author(s):  
Youssef Mourdi ◽  
Mohamed Sadgal ◽  
Hamada El Kabtane ◽  
Wafaa Berrada Fathi

Purpose Even if MOOCs (massive open online courses) are becoming a trend in distance learning, they suffer from a very high rate of learners’ dropout, and as a result, on average, only 10 per cent of enrolled learners manage to obtain their certificates of achievement. This paper aims to give tutors a clearer vision for an effective and personalized intervention as a solution to “retain” each type of learner at risk of dropping out. Design/methodology/approach This paper presents a methodology to provide predictions on learners’ behaviors. This work, which uses a Stanford data set, was divided into several phases, namely, a data extraction, an exploratory study and then a multivariate analysis to reduce dimensionality and to extract the most relevant features. The second step was the comparison between five machine learning algorithms. Finally, the authors used the principle of association rules to extract similarities between the behaviors of learners who dropped out from the MOOC. Findings The results of this work have given that deep learning ensures the best predictions in terms of accuracy, which is an average of 95.8 per cent, and is comparable to other measures such as precision, AUC, Recall and F1 score. Originality/value Many research studies have tried to tackle the MOOC dropout problem by proposing different dropout predictive models. In the same context, comes the present proposal with which the authors have tried to predict not only learners at a risk of dropping out of the MOOCs but also those who will succeed or fail.


2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Burak Cankaya ◽  
Berna Eren Tokgoz ◽  
Ali Dag ◽  
K.C. Santosh

Purpose This paper aims to propose a machine learning-based automatic labeling methodology for chemical tanker activities that can be applied to any port with any number of active tankers and the identification of important predictors. The methodology can be applied to any type of activity tracking that is based on automatically generated geospatial data. Design/methodology/approach The proposed methodology uses three machine learning algorithms (artificial neural networks, support vector machines (SVMs) and random forest) along with information fusion (IF)-based sensitivity analysis to classify chemical tanker activities. The data set is split into training and test data based on vessels, with two vessels in the training data and one in the test data set. Important predictors were identified using a receiver operating characteristic comparative approach, and overall variable importance was calculated using IF from the top models. Findings Results show that an SVM model has the best balance between sensitivity and specificity, at 93.5% and 91.4%, respectively. Speed, acceleration and change in the course on the ground for the vessels are identified as the most important predictors for classifying vessel activity. Research limitations/implications The study evaluates the vessel movements waiting between different terminals in the same port, but not their movements between different ports for their tank-cleaning activities. Practical implications The findings in this study can be used by port authorities, shipping companies, vessel operators and other stakeholders for decision support, performance tracking, as well as for automated alerts. Originality/value This analysis makes original contributions to the existing literature by defining and demonstrating a methodology that can automatically label vehicle activity based on location data and identify certain characteristics of the activity by finding important location-based predictors that effectively classify the activity status.


2020 ◽  
Vol 38 (3) ◽  
pp. 213-225 ◽  
Author(s):  
Agostino Valier

PurposeIn the literature there are numerous tests that compare the accuracy of automated valuation models (AVMs). These models first train themselves with price data and property characteristics, then they are tested by measuring their ability to predict prices. Most of them compare the effectiveness of traditional econometric models against the use of machine learning algorithms. Although the latter seem to offer better performance, there is not yet a complete survey of the literature to confirm the hypothesis.Design/methodology/approachAll tests comparing regression analysis and AVMs machine learning on the same data set have been identified. The scores obtained in terms of accuracy were then compared with each other.FindingsMachine learning models are more accurate than traditional regression analysis in their ability to predict value. Nevertheless, many authors point out as their limit their black box nature and their poor inferential abilities.Practical implicationsAVMs machine learning offers a huge advantage for all real estate operators who know and can use them. Their use in public policy or litigation can be critical.Originality/valueAccording to the author, this is the first systematic review that collects all the articles produced on the subject done comparing the results obtained.


Sign in / Sign up

Export Citation Format

Share Document