Elsevier

Neurocomputing

Volume 73, Issues 7–9, March 2010, Pages 1109-1116
Neurocomputing

Median fuzzy c-means for clustering dissimilarity data

https://doi.org/10.1016/j.neucom.2009.11.020Get rights and content

Abstract

Median clustering is a powerful methodology for prototype based clustering of similarity/dissimilarity data. In this contribution we combine the median c-means algorithm with the fuzzy c-means approach, which is only applicable for vectorial (metric) data in its original variant. For the resulted median fuzzy c-means approach we prove convergence and investigate the behavior of the algorithm in several experiments including real world data from psychotherapy research.

Introduction

Clustering of objects frequently is seen as a partitioning of objects/data into smaller subsets such that those objects are grouped together, which have a similar semantically meaning, characteristics, or behavior. Thereby, the clustering approaches may differ in various aspects like flat or hierarchical structure, crisp or soft (fuzzy) cluster assignments, visualization driven approaches, etc. Yet, the similarity is frequently difficult to capture. For metric data usually the Euclidean metric is applied. However, this favored choice may be inappropriate as it is the case in functional data analysis, for example [44]. Generally, the problem of an adequate chosen similarity is crucially and heavily influencing the clustering result as well as the number of meaningful clusters [24], [47]. A further problem arising in clustering is the choice of an appropriate quality criteria for a given cluster solution. This problem leads to a large number of quality measures for cluster solutions [43]. Prominent examples are the Fishers information relating the intra- and the inter-cluster correlation, the κ statistics based on χ2statistics or Bezdek's cluster index [16].

Basic classic cluster approaches like hierarchical clustering, agglomerative clustering or probabilistic clustering are known from mathematical statistics [46]. Another widely applied paradigm is the prototype based clustering [16]. One basic advantage of prototype based methods in this field, like the famous c-means, is their intuitive understanding by the concept of prototypes as representatives for the clusters [32]. Several prototype based methods have been established ranging from statistical approaches [48] and graph methods [37] to neural vector quantizers [20], [26], [42].

Further, one can distinguish between hard and soft variants of cluster algorithms [4], [13], [20]—the most famous of which is fuzzy c-means [17]. Frequently, respective approaches belong to the class of vector quantizers, requiring the data objects and the prototypes to be embedded in a metric vector space. Popular algorithms are, despite the above mentioned c-means variants, the neural gas [38], deterministic annealed vector quantization [45], information theoretic clustering [36] or topographic quantizers like self-organizing maps [34], or soft topographic vector quantization [25], to name just a few. All these algorithms use similarities between objects and prototypes which are described in terms of (Euclidean) distances for cluster determination [35].

Another class of prototype based cluster algorithms relaxes the assumption of a metric space embedding for the objects to be clustered and the prototypes [28], [12], [33]. This restriction is replaced by the weaker requirement that only similarities/dissimilarities or discrete measures between the data objects are given. Respective algorithms are referred as median or relational clustering, message passing or graph clustering methods [19], [49], [11]. Thereby, an underlying metric is assumed for the given dissimilarities in case of relational data clustering, whereas for median clustering this restriction is dropped. These degrees of freedom lead to constraints for the prototypes. For relational data they are assumed as linear combinations of the data. For median clustering this limitation is further reduced to the minimum demand: prototypes have to be data objects.

Fuzzy clustering of relational data based on fuzzy c-means was proposed by Hathaway et al. [31] inspired by the earlier work of Windham about graded numerical classification of dissimilarity data [54]. This approach was extended by Hathaway and Bezdek for more general dissimilarity data, which have to be, however, at least positive, reflexive and symmetric: In this extension the (non-Euclidean) dissimilarity matrix is transformed by the βspread transform into an Euclidean matrix, such that the clustering is not longer realized on the original data [30]. Other approaches use the Dempster–Shafer theory of belief functions (or evidence theory) to determine a basic belief assignment (or mass function) to each object in such a way that the degree of conflicts between the masses given to any two objects reflects their dissimilarity [14], [39].

In the present contribution we extend the fuzzy c-means for general dissimilarity data based on the idea of median clustering, i.e. we restrict the prototypes to be data points itself. In this way we continue the work of Hathaway and Bezdek. For this purpose we combine fuzzy c-means with methodology introduced for crisp median clustering [11]. We investigate the new algorithm for synthetic as well as real world data and compare it to other fuzzy clustering methods. For this purpose, we use a fuzzy variant of Cohen's-κ for pairwise and the Fleiss’-κ for multiple comparisons.

Section snippets

The median fuzzy c-means algorithm

The median fuzzy c-means algorithm (M-FCM) is based on the standard fuzzy c-means approach (FCM) introduced by Bezdek [4]. It combines this framework with the recently developed methodology for crisp median clustering [11]. Therefore, we briefly reconsider median c-means and fuzzy c-means at first followed by development of its fuzzy counterpart.

Experiments

We perform several exemplary experiments to show the abilities of the new M-FCM. The first experiments are clustering of overlapping Gaussians generating the distance matrix using the Euclidean metric. Subsequently, we apply the algorithm to cluster text fragments of dialogs from psychotherapy sessions.

Conclusion

In the present contribution we introduced a median variant of fuzzy c-means as a prototype based non-hierarchical cluster algorithm. This approach extends earlier variants of fuzzy c-means which were designed to handle relational data. The median variant allows that the given similarities are neither symmetric nor fulfilling any metric requirements as triangle inequality, etc. Only the intuitive assumptions about meaningful similarity are made. The algorithm itself restricts the prototypes to

Tina Geweniger studied Applied Computer Science at the University of Applied Sciences Mittweida, Germany, and examined in 2001. She is a member of the AG Computational Intelligence group at the University Leipzig. Currently she is involved in her Ph.D. studies focusing on clustering, fuzzy classification and text analysis in the medical and biological field.

References (55)

  • J. Bezdek

    Pattern Recognition with Fuzzy Objective Function Algorithms

    (1981)
  • J. Bezdek et al.

    Approximate clustering in very large relational data

    International Journal of Intelligent Systems

    (2006)
  • W. Bucci

    Psychoanalysis and Cognitive Science: A multiple code theory

    (1997)
  • R. Cilibrasi et al.

    Clustering by compression

    IEEE Transactions on Information Theory

    (2005)
  • J. Cohen

    A coefficient of agreement for nominal scales

    Educational and Psychological Measurement

    (1960)
  • J. Cohen

    Weighted chi square: an extension of the kappa method

    Educational and Psychological Measurement

    (1972)
  • B. Conan-Guez, F. Rossi, A.E. Golli, A fast algorithm for the self-organizing map on dissimilarity data, in: M....
  • M. Cottrell et al.

    SOM-based algorithms for qualitative variables

    Neural Networks

    (2004)
  • R. Dav

    Fuzzy shell-clustering and application to circle detection in digital images

    International Journal of General Systems

    (1990)
  • T. Denoeux et al.

    EVCLUS: evidential clustering of proximity data

    IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics

    (2004)
  • R. Duda et al.

    Pattern Classification and Scene Analysis

    (1973)
  • J. Dunn

    A fuzzy relative of the ISODATA process and its use in detecting compact well-separated clusters

    Journal of Cybernetics

    (1973)
  • J. Fleiss

    Statistical Methods for Rates and Proportions

    (1981)
  • B. Frey et al.

    Clustering by message passing between data points

    Science

    (2007)
  • I. Gath et al.

    Unsupervised optimal fuzzy clustering

    IEEE Transactions on Pattern Analysis and Machine Intelligence

    (1989)
  • T. Geweniger, F.-M. Schleif, A. Hasenfuss, B. Hammer, T. Villmann, Comparison of cluster algorithms for the analysis of...
  • T. Geweniger, D. Zühlke, B. Hammer, T. Villmann, Fuzzy variant of affinity propagation in comparison to median fuzzy...
  • Cited by (30)

    • Stable first-arrival picking through adaptive threshold determination and spatial constraint clustering

      2021, Expert Systems with Applications
      Citation Excerpt :

      Fuzzy c-means clustering algorithm (FCM) was proposed by Ruspini (1969), developed by Dunn (1973) and popularized by Bezdek (2013). The FCM algorithm and its numerous variants, such as FCM_S (Pham, 2002), BCFCM (Ahmed, Yamany, Mohamed, Farag, & Moriarty, 2002), KFCM-S1&S2 (Chen & Zhang, 2004), Shadow c-means (Mitra, Pedrycz, & Barman, 2010) and MFCM (Geweniger, Zülke, Hammer, & Villmann, 2010), have been extensively applied in many fields. Among the above variants, FCM_S (Pham, 2002; Verma, Agrawal, & Sharan, 2016) has received widespread attention due to its strong anti-noise ability.

    • A hybrid elicit teaching learning based optimization with fuzzy c-means (ETLBO-FCM) algorithm for data clustering

      2018, Ain Shams Engineering Journal
      Citation Excerpt :

      FCM is a soft partitioned clustering method introduced by Dunn [2,3] and Bezdek [4]. The first fuzzy concept was introduced by Lotfi Zadeh [5] and thereafter, a number of fuzzy algorithms such as RCFCM [6], BCFCM [7], S-FCM [8], KFCM-S1 & S2 [9], MS-FCM [10], SIFCM [11], GKFCM [12], MFCM [13], Recursive FCM [14], Possibilistic FCM [15], GRFCM [16], and Type-2 FCM [17] were implemented in clustering and other application areas. But among all these developed algorithms, FCM is quite popular due to its fuzziness factor in the membership of cluster objects.

    • A concept reduction approach for fuzzy cognitive map models in decision making and management

      2017, Neurocomputing
      Citation Excerpt :

      Several other clustering methods could also be applied, e.g. DBSCAN [26]. The main advantage of this method is that the a priori definition of the number of clusters is not necessary (in contrast to e.g. Fuzzy C-Means [27] or Median Fuzzy C-Means [28]), but it also creates hard clusters. In spite of this, the proposed new method uses fuzzy clustering which makes possible to create overlapping clusters.

    View all citing articles on Scopus

    Tina Geweniger studied Applied Computer Science at the University of Applied Sciences Mittweida, Germany, and examined in 2001. She is a member of the AG Computational Intelligence group at the University Leipzig. Currently she is involved in her Ph.D. studies focusing on clustering, fuzzy classification and text analysis in the medical and biological field.

    Dietlind Zühlke studied Computer Science in Leipzig and Bonn and finished her diploma thesis in 2006 at the Fraunhofer Institute FIT in Sankt Augustin (Germany). Starting in 2007, she led different projects in the Life Science Informatics Group of the Fraunhofer FIT focusing on image processing and pattern recognition in the biological and medical domain. Since 2009 she has a shared position in cooperation of Fraunhofer FIT and the RWTH Aachen. Her research area comprises machine learning approaches like neural maps, clustering and classification with focus on practical applications in medicine, biology and other life sciences.

    Barbara Hammer received her Ph.D. in Computer Science in 1995 and her venia legendi in Computer Science in 2003, both from the University of Osnabrueck, Germany. From 2000 to 2004, she was leader of the junior research group ‘Learning with neural methods in structured domains’ before accepting an offer as professor for Theoretical Computer Science at Clausthal University of Technology. Her areas of expertise include clustering and classification, recurrent and structure processing neural network, self-organization, statistical learning theory, and bioinformatics.

    Thomas Villmann received his Ph.D. in Computer Science in 1996 and his venia legendi in the same subject in 2005, both from University of Leipzig. From 1997 to 2009 he led the research group of computational intelligence of the clinic for psychotherapy at Leipzig University. Since 2009 he is a professor for Technomathematics and Computational Intelligence at the University of Applied Science Mittweida, Germany. He is a founding member of the German chapter of ENNS. His research areas include a broad range of machine learning approaches like neural maps, clustering, classification, pattern recognition and evolutionary algorithms as well as applications in medicine, bioinformatics, satellite remote sensing and others.

    View full text