Probabilistic Belief Contraction: Considerations on Epistemic Entrenchment, Probability Mixtures and KL Divergence

Author(s):  
Kinzang Chhogyal ◽  
Abhaya Nayak ◽  
Abdul Sattar
Author(s):  
R. R. Gharieb ◽  
G. Gendy ◽  
H. Selim

In this paper, the standard hard C-means (HCM) clustering approach to image segmentation is modified by incorporating weighted membership Kullback–Leibler (KL) divergence and local data information into the HCM objective function. The membership KL divergence, used for fuzzification, measures the proximity between each cluster membership function of a pixel and the locally-smoothed value of the membership in the pixel vicinity. The fuzzification weight is a function of the pixel to cluster-centers distances. The used pixel to a cluster-center distance is composed of the original pixel data distance plus a fraction of the distance generated from the locally-smoothed pixel data. It is shown that the obtained membership function of a pixel is proportional to the locally-smoothed membership function of this pixel multiplied by an exponentially distributed function of the minus pixel distance relative to the minimum distance provided by the nearest cluster-center to the pixel. Therefore, since incorporating the locally-smoothed membership and data information in addition to the relative distance, which is more tolerant to additive noise than the absolute distance, the proposed algorithm has a threefold noise-handling process. The presented algorithm, named local data and membership KL divergence based fuzzy C-means (LDMKLFCM), is tested by synthetic and real-world noisy images and its results are compared with those of several FCM-based clustering algorithms.


Complexity ◽  
2021 ◽  
Vol 2021 ◽  
pp. 1-8
Author(s):  
Lili Wang ◽  
Ting Shi ◽  
Shijin Li

Since the user recommendation complex matrix is characterized by strong sparsity, it is difficult to correctly recommend relevant services for users by using the recommendation method based on location and collaborative filtering. The similarity measure between users is low. This paper proposes a fusion method based on KL divergence and cosine similarity. KL divergence and cosine similarity have advantages by comparing three similar metrics at different K values. Using the fusion method of the two, the user’s similarity with the preference is reused. By comparing the location-based collaborative filtering (LCF) algorithm, user-based collaborative filtering (UCF) algorithm, and user recommendation algorithm (F2F), the proposed method has the preparation rate, recall rate, and experimental effect advantage. In different median values, the proposed method also has an advantage in experimental results.


Sign in / Sign up

Export Citation Format

Share Document