Skip to main content

2013 | OriginalPaper | Buchkapitel

16. Learning Mixtures by Simplifying Kernel Density Estimators

verfasst von : Olivier Schwander, Frank Nielsen

Erschienen in: Matrix Information Geometry

Verlag: Springer Berlin Heidelberg

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Gaussian mixture models are a widespread tool for modeling various and complex probability density functions. They can be estimated by various means, often using Expectation–Maximization or Kernel Density Estimation. In addition to these well known algorithms, new and promising stochastic modeling methods include Dirichlet Process mixtures and k-Maximum Likelihood Estimators. Most of the methods, including Expectation–Maximization, lead to compact models but may be expensive to compute. On the other hand Kernel Density Estimation yields to large models which are computationally cheap to build. In this chapter we present new methods to get high-quality models that are both compact and fast to compute. This is accomplished by the simplification of Kernel Density Estimator. The simplification is a clustering method based on k-means-like algorithms. Like all k-means algorithms, our method rely on divergences and centroids computation and we use two different divergences (and their associated centroids), Bregman and . Along with the description of the algorithms, we describe the pyMEF =library=, which is a Python library designed for the manipulation of mixture of exponential families. Unlike most of the other existing tools, this library allows to use any exponential family instead of being limited to a particular distribution. The generic library allows to rapidly explore the different available exponential families in order to choose the better suited for a particular application. We evaluate the proposed algorithms by building mixture models on examples from a bio-informatics application. The quality of the resulting models is measured in terms of log-likelihood and of Kullback–Leibler divergence.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
The product of exponential families is an exponential family.
 
Literatur
1.
Zurück zum Zitat Banerjee, A., Merugu, S., Dhillon, I.S., Ghosh, J.: Clustering with Bregman divergences. J. Mach. Learn. Res. 6, 1705–1749 (2005)MathSciNetMATH Banerjee, A., Merugu, S., Dhillon, I.S., Ghosh, J.: Clustering with Bregman divergences. J. Mach. Learn. Res. 6, 1705–1749 (2005)MathSciNetMATH
2.
Zurück zum Zitat Bernauer, J., Huang, X., Sim, A.Y.L., Levitt, M.: Fully differentiable coarse-grained and all-atom knowledge-based potentials for RNA structure evaluation. RNA 17(6), 1066 (2011)CrossRef Bernauer, J., Huang, X., Sim, A.Y.L., Levitt, M.: Fully differentiable coarse-grained and all-atom knowledge-based potentials for RNA structure evaluation. RNA 17(6), 1066 (2011)CrossRef
3.
Zurück zum Zitat Biernacki, C., Celeux, G., Govaert, G., Langrognet, F.: Model-based cluster and discriminant analysis with the MIXMOD software. Comput. Stat. Data Anal. 51(2), 587–600 (2006)MathSciNetMATHCrossRef Biernacki, C., Celeux, G., Govaert, G., Langrognet, F.: Model-based cluster and discriminant analysis with the MIXMOD software. Comput. Stat. Data Anal. 51(2), 587–600 (2006)MathSciNetMATHCrossRef
4.
Zurück zum Zitat Brown, L.D.: Fundamentals of statistical exponential families: with applications in statistical decision theory. IMS (1986) Brown, L.D.: Fundamentals of statistical exponential families: with applications in statistical decision theory. IMS (1986)
5.
Zurück zum Zitat Čencov, N.N.: Statistical Decision Rules and Optimal Inference, Translations of Mathematical Monographs, vol. 53. American Mathematical Society, Providence, R.I. (1982). (Translation from the Russian edited by Lev J. Leifman) Čencov, N.N.: Statistical Decision Rules and Optimal Inference, Translations of Mathematical Monographs, vol. 53. American Mathematical Society, Providence, R.I. (1982). (Translation from the Russian edited by Lev J. Leifman)
6.
Zurück zum Zitat Costa, S.I.R., Santos, S.A., Strapasson, J.E.: Fisher information matrix and hyperbolic geometry. In: Information Theory Workshop, 2005 IEEE, 3 pp, 29 Aug-1 Sept (2005) Costa, S.I.R., Santos, S.A., Strapasson, J.E.: Fisher information matrix and hyperbolic geometry. In: Information Theory Workshop, 2005 IEEE, 3 pp, 29 Aug-1 Sept (2005)
7.
Zurück zum Zitat Dempster, A.P., Laird, N.M., Rubin, D.B.: Maximum likelihood from incomplete data via the em algorithm. J. R. Stat. Soc. Ser. B (Methodological), 1–38 (1977) Dempster, A.P., Laird, N.M., Rubin, D.B.: Maximum likelihood from incomplete data via the em algorithm. J. R. Stat. Soc. Ser. B (Methodological), 1–38 (1977)
8.
Zurück zum Zitat Galperin, G.A.: A concept of the mass center of a system of material points in the constant curvature spaces. Commun. Math. Phys. 154(1), 63–84 (1993)MathSciNetMATHCrossRef Galperin, G.A.: A concept of the mass center of a system of material points in the constant curvature spaces. Commun. Math. Phys. 154(1), 63–84 (1993)MathSciNetMATHCrossRef
9.
Zurück zum Zitat Garcia, V., Nielsen, F., Nock, R.: Levels of details for gaussian mixture models. In: Computer Vision-ACCV 2009, 514–525 (2010) Garcia, V., Nielsen, F., Nock, R.: Levels of details for gaussian mixture models. In: Computer Vision-ACCV 2009, 514–525 (2010)
10.
Zurück zum Zitat Georgi, B., Costa, I.G., Schliep, A.: PyMix–the Python mixture package–a tool for clustering of heterogeneous biological data. BMC Bioinf. 11(1), 9 (2010)CrossRef Georgi, B., Costa, I.G., Schliep, A.: PyMix–the Python mixture package–a tool for clustering of heterogeneous biological data. BMC Bioinf. 11(1), 9 (2010)CrossRef
11.
Zurück zum Zitat Ji, Y., Wu, C., Liu, P., Wang, J., Coombes, K.R.: Applications of beta-mixture models in bioinformatics. Bioinformatics 21(9), 2118 (2005)CrossRef Ji, Y., Wu, C., Liu, P., Wang, J., Coombes, K.R.: Applications of beta-mixture models in bioinformatics. Bioinformatics 21(9), 2118 (2005)CrossRef
12.
Zurück zum Zitat Kass, R.E., Vos, P.W.: Geometrical Foundations of Asymptotic Inference. Wiley, New York (1987) Kass, R.E., Vos, P.W.: Geometrical Foundations of Asymptotic Inference. Wiley, New York (1987)
13.
Zurück zum Zitat Mayrose, I. Friedman, N. Pupko, T.: A gamma mixture model better accounts for among site rate heterogeneity. Bioinformatics 21(Suppl 2), ii151-ii158 (2005) Mayrose, I. Friedman, N. Pupko, T.: A gamma mixture model better accounts for among site rate heterogeneity. Bioinformatics 21(Suppl 2), ii151-ii158 (2005)
14.
Zurück zum Zitat Nielsen, F. Boltz, S. Schwander, O.: Bhattacharyya clustering with applications to mixture simplifications. In: IEEE International Conference on Pattern Recognition, Istanbul, Turkey, ICPR’10 (2010) Nielsen, F. Boltz, S. Schwander, O.: Bhattacharyya clustering with applications to mixture simplifications. In: IEEE International Conference on Pattern Recognition, Istanbul, Turkey, ICPR’10 (2010)
15.
Zurück zum Zitat Nielsen, F., Garcia. V.: Statistical exponential families: a digest with flash cards. arXiv:0911.4863 (2009) Nielsen, F., Garcia. V.: Statistical exponential families: a digest with flash cards. arXiv:0911.4863 (2009)
16.
Zurück zum Zitat Nielsen, F., Nock, R.: Hyperbolic voronoi diagrams made easy. arXiv:0903.3287 (2009) Nielsen, F., Nock, R.: Hyperbolic voronoi diagrams made easy. arXiv:0903.3287 (2009)
17.
Zurück zum Zitat Nielsen, F., Nock. R.: Jensen-bregman voronoi diagrams and centroidal tessellations. In: 2010 International Symposium on Voronoi Diagrams in Science and Engineering (ISVD), pp. 56–65. IEEE (2010) Nielsen, F., Nock. R.: Jensen-bregman voronoi diagrams and centroidal tessellations. In: 2010 International Symposium on Voronoi Diagrams in Science and Engineering (ISVD), pp. 56–65. IEEE (2010)
20.
Zurück zum Zitat Rasmussen, C.E.: The infinite gaussian mixture model. Adv. Neural Inf. Process. Systems 12, 554–560 (2000) Rasmussen, C.E.: The infinite gaussian mixture model. Adv. Neural Inf. Process. Systems 12, 554–560 (2000)
21.
Zurück zum Zitat Reverter, F., Oller, J.M.: Computing the rao distance for gamma distributions. J. Comput. Appl. Math. 157(1), 155–167 (2003)MathSciNetMATHCrossRef Reverter, F., Oller, J.M.: Computing the rao distance for gamma distributions. J. Comput. Appl. Math. 157(1), 155–167 (2003)MathSciNetMATHCrossRef
22.
Zurück zum Zitat Rong, G., Jin, M., Guo, X.: Hyperbolic centroidal voronoi tessellation. In: Proceedings of the 14th ACM Symposium on Solid and Physical Modeling, SPM ’10, pp. 117–126. ACM, New York, NY, USA (2010) Rong, G., Jin, M., Guo, X.: Hyperbolic centroidal voronoi tessellation. In: Proceedings of the 14th ACM Symposium on Solid and Physical Modeling, SPM ’10, pp. 117–126. ACM, New York, NY, USA (2010)
23.
Zurück zum Zitat Schwander, O., Nielsen, F.: Model centroids for the simplification of kernel density estimators. In: 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), March (2012) Schwander, O., Nielsen, F.: Model centroids for the simplification of kernel density estimators. In: 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), March (2012)
24.
Zurück zum Zitat Seabra, J.C., Ciompi, F., Pujol, O., Mauri, J., Radeva, P., Sanches, J.: Rayleigh mixture model for plaque characterization in intravascular ultrasound. IEEE Trans. Biomed. Eng. 58(5), 1314–1324 (2011)CrossRef Seabra, J.C., Ciompi, F., Pujol, O., Mauri, J., Radeva, P., Sanches, J.: Rayleigh mixture model for plaque characterization in intravascular ultrasound. IEEE Trans. Biomed. Eng. 58(5), 1314–1324 (2011)CrossRef
25.
Zurück zum Zitat Sheather, S.J., Jones, M.C.: A reliable data-based bandwidth selection method for kernel density estimation. J. R. Stat. Soc. Ser. B (Methodological) 53(3), 683–690 (1991)MathSciNetMATH Sheather, S.J., Jones, M.C.: A reliable data-based bandwidth selection method for kernel density estimation. J. R. Stat. Soc. Ser. B (Methodological) 53(3), 683–690 (1991)MathSciNetMATH
26.
Zurück zum Zitat Sim, A.Y.L., Schwander, O., Levitt, M., Bernauer, J.: Evaluating mixture models for building rna knowledge-based potentials. J. Bioinf. Comput. Biol. (2012) Sim, A.Y.L., Schwander, O., Levitt, M., Bernauer, J.: Evaluating mixture models for building rna knowledge-based potentials. J. Bioinf. Comput. Biol. (2012)
Metadaten
Titel
Learning Mixtures by Simplifying Kernel Density Estimators
verfasst von
Olivier Schwander
Frank Nielsen
Copyright-Jahr
2013
Verlag
Springer Berlin Heidelberg
DOI
https://doi.org/10.1007/978-3-642-30232-9_16

Neuer Inhalt