Skip to main content

2016 | OriginalPaper | Buchkapitel

Learning Latent Features with Infinite Non-negative Binary Matrix Tri-factorization

verfasst von : Xi Yang, Kaizhu Huang, Rui Zhang, Amir Hussain

Erschienen in: Neural Information Processing

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Non-negative Matrix Factorization (NMF) has been widely exploited to learn latent features from data. However, previous NMF models often assume a fixed number of features, say p features, where p is simply searched by experiments. Moreover, it is even difficult to learn binary features, since binary matrix involves more challenging optimization problems. In this paper, we propose a new Bayesian model called infinite non-negative binary matrix tri-factorizations model (iNBMT), capable of learning automatically the latent binary features as well as feature number based on Indian Buffet Process (IBP). Moreover, iNBMT engages a tri-factorization process that decomposes a nonnegative matrix into the product of three components including two binary matrices and a non-negative real matrix. Compared with traditional bi-factorization, the tri-factorization can better reveal the latent structures among items (samples) and attributes (features). Specifically, we impose an IBP prior on the two infinite binary matrices while a truncated Gaussian distribution is assumed on the weight matrix. To optimize the model, we develop an efficient modified maximization-expectation algorithm (ME-algorithm), with the iteration complexity one order lower than another recently-proposed Maximization-Expectation-IBP model [9]. We present the model definition, detail the optimization, and finally conduct a series of experiments. Experimental results demonstrate that our proposed iNBMT model significantly outperforms the other comparison algorithms in both synthetic and real data.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
Since IBP-IBP is mainly for clustering, we do not show its (almost messy) reconstruction results for fairness.
 
Literatur
1.
Zurück zum Zitat Ding, C., He, X., Simon, H.D.: On the equivalence of nonnegative matrix factorization and spectral clustering. In: SIAM International Conference on Data Mining (2005) Ding, C., He, X., Simon, H.D.: On the equivalence of nonnegative matrix factorization and spectral clustering. In: SIAM International Conference on Data Mining (2005)
2.
Zurück zum Zitat Ding, C., Li, T., Peng, W., Park, H.: Orthogonal nonnegative matrix tri-factorizations for clustering. In: Proceedings of the Twelfth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 126–135. ACM Press (2006) Ding, C., Li, T., Peng, W., Park, H.: Orthogonal nonnegative matrix tri-factorizations for clustering. In: Proceedings of the Twelfth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 126–135. ACM Press (2006)
3.
Zurück zum Zitat Doshi-velez, F., Miller, K.T., Van Gael, J., Teh, Y.W.: Variational inference for the Indian buffet process. In: Proceedings of the Twelfth International Conference on Artificial Intelligence and Statistics, AISTATS, pp. 137–144 (2009) Doshi-velez, F., Miller, K.T., Van Gael, J., Teh, Y.W.: Variational inference for the Indian buffet process. In: Proceedings of the Twelfth International Conference on Artificial Intelligence and Statistics, AISTATS, pp. 137–144 (2009)
4.
Zurück zum Zitat Ghahramani, Z., Beal, M.J.: Propagation algorithms for variational Bayesian learning. In: Advances in Neural Information Processing Systems, vol. 13, pp. 507–513 (2000) Ghahramani, Z., Beal, M.J.: Propagation algorithms for variational Bayesian learning. In: Advances in Neural Information Processing Systems, vol. 13, pp. 507–513 (2000)
5.
Zurück zum Zitat Griffiths, T.L., Ghahramani, Z.: Infinite latent feature models and the Indian buffet process. In: Advances in Neural Information Processing Systems, vol. 18, pp. 475–482 (2005) Griffiths, T.L., Ghahramani, Z.: Infinite latent feature models and the Indian buffet process. In: Advances in Neural Information Processing Systems, vol. 18, pp. 475–482 (2005)
6.
Zurück zum Zitat Knowles, D., Ghahramani, Z.: Infinite sparse factor analysis and infinite independent components analysis. In: Davies, M.E., James, C.J., Abdallah, S.A., Plumbley, M.D. (eds.) ICA 2007. LNCS, vol. 4666, pp. 381–388. Springer, Heidelberg (2007)CrossRef Knowles, D., Ghahramani, Z.: Infinite sparse factor analysis and infinite independent components analysis. In: Davies, M.E., James, C.J., Abdallah, S.A., Plumbley, M.D. (eds.) ICA 2007. LNCS, vol. 4666, pp. 381–388. Springer, Heidelberg (2007)CrossRef
7.
Zurück zum Zitat Kurihara, K., Welling, M.: Bayesian \({k}\text{- }\text{ means }\) as a “maximization-expectation” algorithm. Neural Comput. 21(4), 1145–1172 (2009)MathSciNetCrossRefMATH Kurihara, K., Welling, M.: Bayesian \({k}\text{- }\text{ means }\) as a “maximization-expectation” algorithm. Neural Comput. 21(4), 1145–1172 (2009)MathSciNetCrossRefMATH
8.
Zurück zum Zitat Lee, D.D., Seung, H.S.: Algorithms for non-negative matrix factorization. In: NIPS, pp. 556–562. MIT Press (2001) Lee, D.D., Seung, H.S.: Algorithms for non-negative matrix factorization. In: NIPS, pp. 556–562. MIT Press (2001)
9.
Zurück zum Zitat Reed, C., Ghahramani, Z.: Scaling the Indian buffet process via submodular maximization. In: Proceedings of the 30th International Conference on Machine Learning, pp. 1013–1021(2013) Reed, C., Ghahramani, Z.: Scaling the Indian buffet process via submodular maximization. In: Proceedings of the 30th International Conference on Machine Learning, pp. 1013–1021(2013)
10.
Zurück zum Zitat Zhang, Z., Li, T., Ding, C.H.Q., Ren, X.-W., Zhang, X.-S.: Binary matrix factorization for analyzing gene expression data. Data Min. Knowl. Discov. 20(1), 28–52 (2010)MathSciNetCrossRef Zhang, Z., Li, T., Ding, C.H.Q., Ren, X.-W., Zhang, X.-S.: Binary matrix factorization for analyzing gene expression data. Data Min. Knowl. Discov. 20(1), 28–52 (2010)MathSciNetCrossRef
Metadaten
Titel
Learning Latent Features with Infinite Non-negative Binary Matrix Tri-factorization
verfasst von
Xi Yang
Kaizhu Huang
Rui Zhang
Amir Hussain
Copyright-Jahr
2016
DOI
https://doi.org/10.1007/978-3-319-46687-3_65