On Objective-Based Rough Clustering with Fuzzy-Set Representation

Author(s):  
Naohiko Kinoshita ◽  
◽  
Yasunori Endo ◽  
Ken Onishi ◽  
◽  
...  

The rough clustering algorithm we proposed based on the optimization of objective function (RCM) has a problem because conventional rough clustering algorithm results do not ensure that solutions are optimal. To solve this problem, we propose rough clustering algorithms based on optimization of an objective function with fuzzy-set representation. This yields more flexible results than RCM. We verify algorithm effectiveness through numerical examples.

Author(s):  
Makito Yamashiro ◽  
◽  
Yasunori Endo ◽  
Yukihiro Hamasuna ◽  

The clustering algorithm we propose is based on probabilistic dissimilarity, which is formed by introducing the concept of probability into conventional dissimilarity. After defining probabilistic dissimilarity, we present examples of probabilistic dissimilarity functions. After considering an objective function with probabilistic dissimilarity. Furthermore, we construct a clustering algorithm probabilistic dissimilarity based using optimal solutions maximizing the objective function. Numerical examples verify the effectiveness of our algorithm.


Author(s):  
Yukihiro Hamasuna ◽  
◽  
Yasunori Endo ◽  
Sadaaki Miyamoto ◽  

Detecting various kinds of cluster shape is an important problem in the field of clustering. In general, it is difficult to obtain clusters with different sizes or shapes by single-objective function. From that sense, we have proposed the concept of clusterwise tolerance and constructed clustering algorithms based on it. In the field of data mining, regularization techniques are used in order to derive significant classifiers. In this paper, we propose another concept of clusterwise tolerance from the viewpoint of regularization. Moreover, we construct clustering algorithms for data with clusterwise tolerance based onL2- andL1-regularization. After that, we describe fuzzy classification functions of proposed algorithms. Finally, we show the effectiveness of proposed algorithms through numerical examples.


Author(s):  
Yasunori Endo ◽  
◽  
Arisa Taniguchi ◽  
Yukihiro Hamasuna ◽  
◽  
...  

Clustering is an unsupervised classification technique for data analysis. In general, each datum in real space is transformed into a point in a pattern space to apply clustering methods. Data cannot often be represented by a point, however, because of its uncertainty, e.g., measurement error margin and missing values in data. In this paper, we will introduce quadratic penalty-vector regularization to handle such uncertain data using Hard c-Means (HCM), which is one of the most typical clustering algorithms. We first propose a new clustering algorithm called hard c-means using quadratic penalty-vector regularization for uncertain data (HCMP). Second, we propose sequential extraction hard c-means using quadratic penalty-vector regularization (SHCMP) to handle datasets whose cluster number is unknown. Furthermore, we verify the effectiveness of our proposed algorithms through numerical examples.


2010 ◽  
Vol 44-47 ◽  
pp. 3897-3901
Author(s):  
Hsiang Chuan Liu ◽  
Yen Kuei Yu ◽  
Jeng Ming Yih ◽  
Chin Chun Chen

Euclidean distance function based fuzzy clustering algorithms can only be used to detect spherical structural clusters. Gustafson-Kessel (GK) clustering algorithm and Gath-Geva (GG) clustering algorithm were developed to detect non-spherical structural clusters by employing Mahalanobis distance in objective function, however, both of them need to add some constrains for Mahalanobis distance. In this paper, the authors’ improved Fuzzy C-Means algorithm based on common Mahalanobis distance (FCM-CM) is used to identify the mastery concepts in linear algebra, for comparing the performances with other four partition algorithms; FCM-M, GG, GK, and FCM. The result shows that FCM-CM has better performance than others.


Author(s):  
Naohiko Kinoshita ◽  
◽  
Yasunori Endo ◽  

Clustering is one of the most popular unsupervised classification methods. In this paper, we focus on rough clustering methods based on rough-set representation. Rough k-Means (RKM) is one of the rough clustering method proposed by Lingras et al. Outputs of many clustering algorithms, including RKM depend strongly on initial values, so we must evaluate the validity of outputs. In the case of objectivebased clustering algorithms, the objective function is handled as the measure. It is difficult, however to evaluate the output in RKM, which is not objective-based. To solve this problem, we propose new objective-based rough clustering algorithms and verify theirs usefulness through numerical examples.


Mathematics ◽  
2019 ◽  
Vol 7 (1) ◽  
pp. 36 ◽  
Author(s):  
Jiongmei Mo ◽  
Han-Liang Huang

Fuzzy clustering is widely used in business, biology, geography, coding for the internet and more. A single-valued neutrosophic set is a generalized fuzzy set, and its clustering algorithm has attracted more and more attention. An equivalence matrix is a common tool in clustering algorithms. At present, there exist no results constructing a single-valued neutrosophic number equivalence matrix using t-norm and t-conorm. First, the concept of a ( T , S ) -based composition matrix is defined in this paper, where ( T , S ) is a dual pair of triangular modules. Then, a ( T , S ) -based single-valued neutrosophic number equivalence matrix is given. A λ -cutting matrix of single-valued neutrosophic number matrix is also introduced. Moreover, their related properties are studied. Finally, an example and comparison experiment are given to illustrate the effectiveness and superiority of our proposed clustering algorithm.


2020 ◽  
Vol 24 (5) ◽  
pp. 977-992
Author(s):  
Yue Liu ◽  
Bufang Li

Clustering algorithm is the foundation and important technology in data mining. In fact, in the real world, the data itself often has a hierarchical structure. Hierarchical clustering aims at constructing a cluster tree, which reveals the underlying modal structure of a complex density. Due to its inherent complexity, most existing hierarchical clustering algorithms are usually designed heuristically without an explicit objective function, which limits its utilization and analysis. K-means clustering, the well-known simple yet effective algorithm which can be expressed from the view of probability distribution, has inherent connection to Mixture of Gaussians (MoG). At this point, we consider combining Bayesian theory analysis with K-means algorithm. This motivates us to develop a hierarchical clustering based on K-means under the probability distribution framework, which is different from existing hierarchical K-means algorithms processing data in a single-pass manner along with heuristic strategies. For this goal, we propose an explicit objective function for hierarchical clustering, termed as Bayesian hierarchical K-means (BHK-means). In our method, a cascaded clustering tree is constructed, in which all layers interact with each other in the network-like manner. In this cluster tree, the clustering results of each layer are influenced by the parent and child nodes. Therefore, the clustering result of each layer is dynamically improved in accordance with the global hierarchical clustering objective function. The objective function is solved using the same algorithm as K-means, the Expectation-maximization algorithm. The experimental results on both synthetic data and benchmark datasets demonstrate the effectiveness of our algorithm over the existing related ones.


Author(s):  
Yasunori Endo ◽  
◽  
Ayako Heki ◽  
Yukihiro Hamasuna ◽  
◽  
...  

The non metricmodel is a kind of clustering method in which belongingness or the membership grade of each object in each cluster is calculated directly from dissimilarities between objects and in which cluster centers are not used. The clustering field has recently begun to focus on rough set representation instead of fuzzy set representation. Conventional clustering algorithms classify a set of objects into clusters with clear boundaries, that is, one object must belong to one cluster. Many objects in the real world, however, belong to more than one cluster because cluster boundaries overlap each other. Fuzzy set representation of clusters makes it possible for each object to belong to more than one cluster. The fuzzy degree of membership may, however, be too descriptive for interpreting clustering results. Rough set representation handles such cases. Clustering based on rough sets could provide a solution that is less restrictive than conventional clustering and more descriptive than fuzzy clustering. This paper covers two types of Rough-set-based Non Metric model (RNM). One algorithm is the Roughset-based Hard Non Metric model (RHNM) and the other is the Rough-set-based Fuzzy Non Metric model (RFNM). In both algorithms, clusters are represented by rough sets and each cluster consists of lower and upper approximation. The effectiveness of proposed algorithms is evaluated through numerical examples.


Author(s):  
Yuchi Kanzawa ◽  

In this paper, two types of fuzzy co-clustering algorithms are proposed. First, it is shown that the base of the objective function for the conventional fuzzy co-clustering method is very similar to the base for entropy-regularized fuzzy nonmetric model. Next, it is shown that the non-sense clustering problem in the conventional fuzzy co-clustering algorithms is identical to that in fuzzy nonmetric model algorithms, in the case that all dissimilarities among rows and columns are zero. Based on this discussion, a method is proposed applying entropy-regularized fuzzy nonmetric model after all dissimilarities among rows and columns are set to some values using a TIBA imputation technique. Furthermore, since relational fuzzy cmeans is similar to fuzzy nonmetricmodel, in the sense that both methods are designed for homogeneous relational data, a method is proposed applying entropyregularized relational fuzzyc-means after imputing all dissimilarities among rows and columns with TIBA. Some numerical examples are presented for the proposed methods.


Author(s):  
Mohana Priya K ◽  
Pooja Ragavi S ◽  
Krishna Priya G

Clustering is the process of grouping objects into subsets that have meaning in the context of a particular problem. It does not rely on predefined classes. It is referred to as an unsupervised learning method because no information is provided about the "right answer" for any of the objects. Many clustering algorithms have been proposed and are used based on different applications. Sentence clustering is one of best clustering technique. Hierarchical Clustering Algorithm is applied for multiple levels for accuracy. For tagging purpose POS tagger, porter stemmer is used. WordNet dictionary is utilized for determining the similarity by invoking the Jiang Conrath and Cosine similarity measure. Grouping is performed with respect to the highest similarity measure value with a mean threshold. This paper incorporates many parameters for finding similarity between words. In order to identify the disambiguated words, the sense identification is performed for the adjectives and comparison is performed. semcor and machine learning datasets are employed. On comparing with previous results for WSD, our work has improvised a lot which gives a percentage of 91.2%


Sign in / Sign up

Export Citation Format

Share Document