scholarly journals Learning Naïve Bayes Tree for Conditional Probability Estimation

Author(s):  
Han Liang ◽  
Yuhong Yan
Author(s):  
Sikha Bagui ◽  
Keerthi Devulapalli ◽  
Sharon John

This study presents an efficient way to deal with discrete as well as continuous values in Big Data in a parallel Naïve Bayes implementation on Hadoop's MapReduce environment. Two approaches were taken: (i) discretizing continuous values using a binning method; and (ii) using a multinomial distribution for probability estimation of discrete values and a Gaussian distribution for probability estimation of continuous values. The models were analyzed and compared for performance with respect to run time and classification accuracy for varying data sizes, data block sizes, and map memory sizes.


2012 ◽  
Vol 26 ◽  
pp. 239-245 ◽  
Author(s):  
Liangxiao Jiang ◽  
Zhihua Cai ◽  
Dianhong Wang ◽  
Harry Zhang

This paper proposes an effective concept of mining the feedback of product given by the user. In return various solutions are suggested according to the ratings of the aspect and its corresponding weightage. The satisfaction of user is determined by the help of user’s rating and weight of the aspect determines the significance of each aspect in the user’s review. These methodologies are thus, important and play a significant role for the manufacturers and producers to improvise their product and eventually leading to rise in the market value of that particular product. The methodology here extracts the aspects from the feedbacks of users with the help of conditional probability and bootstrap technique. Also an approach that is supervised and is called by the name, Naïve Bayes is used to classify aspect ratings and the sentiment words are considered as properties or features.


Author(s):  
HAN LIANG ◽  
YUHONG YAN ◽  
HARRY ZHANG

In machine learning and data mining, traditional learning models aim for high classification accuracy. However, accurate class probability prediction is more desirable than classification accuracy in many practical applications, such as medical diagnosis. Although it is known that decision trees can be adapted to be class probability estimators in a variety of approaches, and the resulting models are uniformly called Probability Estimation Trees (PETs), the performances of these PETs in class probability estimation, have not yet been investigated. We begin our research by empirically studying PETs in terms of class probability estimation, measured by Log Conditional Likelihood (LCL). We also compare a PET called C4.4 with other representative models, including Naïve Bayes, Naïve Bayes Tree, Bayesian Network, KNN and SVM, in LCL. From our experiments, we draw several valuable conclusions. First, among various tree-based models, C4.4 is the best in yielding precise class probability prediction measured by LCL. We provide an explanation for this and reveal the nature of LCL. Second, compared with non tree-based models, C4.4 also performs best. Finally, LCL does not dominate another well-established relevant metric — AUC, which suggests that different decision-tree learning models should be used for different objectives. Our experiments are conducted on the basis of 36 UCI sample sets. We run all the models within a machine learning platform — Weka. We also explore an approach to improve the class probability estimation of Naïve Bayes Tree. We propose a greedy and recursive learning algorithm, where at each step, LCL is used as the scoring function to expand the decision tree. The algorithm uses Naïve Bayes created at leaves to estimate class probabilities of test samples. The whole tree encodes the posterior class probability in its structure. One benefit of improving class probability estimation is that both classification accuracy and AUC can be possibly scaled up. We call the new model LCL Tree (LCLT). Our experiments on 33 UCI sample sets show that LCLT outperforms all state-of-the-art learning models, such as Naïve Bayes Tree, significantly in accurate class probability prediction measured by LCL, as well as in classification accuracy and AUC.


2020 ◽  
Vol 7 (1) ◽  
pp. 46-54
Author(s):  
Jasman Pardede

Pesatnya perkembangan teknologi dan media sosial dapat memudahkan pengguna untuk menyampaikan informasi. Selain itu, media sosial juga memberikan dampak negatif dengan cara memposting tulisan kejam atau berkomentar semena-mena tanpa memikirkan akibat pada orang lain. Hal inilah yang menjadikan salah satu terjadinya tindak kekerasan dalam dunia maya (Cyberbullying). Tahapan awal yang dilakukan dalam penelitian ini adalah pengolahan bahasa atau yang disebut dengan text preprocessing meliputi tokenizing,casefolding, stopword removal dan stemming. Kemudian feature selection yaitu mengubah dokument teks menjadi matriks dengan tujuan untuk mendapatkan fitur pada setiap kata untuk dijadikan parameter atau kriteria klasifikasi. Untuk pengambilan keputusan apakah komentar mengandung makna bully atau nonbully menggunakan algoritma Naïve Bayes Classification dengan model multinomial naïve bayes. Perhitungan yang dilakukan adalah menghitung nilai probabilitas setiap kata yang muncul berdasarkan classdan nilai perkalian class conditional probability. Berdasarkan hasil eksperimen menggunakan dataset “cyberbullying comments” yang diambil dari Kaggle  akurasi yang didapat sebesar 80%, precission 81% dan recall 80%.


Sign in / Sign up

Export Citation Format

Share Document