Skip to main content
Log in

A hybrid language model based on a recurrent neural network and probabilistic topic modeling

  • Applied Problems
  • Published:
Pattern Recognition and Image Analysis Aims and scope Submit manuscript

Abstract

A language model based on features extracted from a recurrent neural network language model and semantic embedding of the left context of the current word based on probabilistic semantic analysis (PLSA) is developed. To calculate such embedding, the context is considered as a document. The effect of vanishing gradients in a recurrent neural network is reduced by this method. The experiment has shown that adding topic-based features reduces perplexity by 10%.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. I. Oparin, “Language models for automatic speech recognition of inflectional languages,” PhD Thesis (Univ. of West Bohemia, Pilsen, 2008).

    Google Scholar 

  2. E. W. D. Whittaker, “Statistical language modeling for automatic speech recognition of Russian and English,” PhD Thesis (Cambridge Univ., 2000).

    Google Scholar 

  3. A. Deoras, T. Mikolov, and S. Kombrik, “Approximate inference: a sampling based modeling technique to capture complex dependencies in a language model,” Speech Commun. (2012).

    Google Scholar 

  4. T. Mikolov, “Statistical language models based on neural networks,” PhD Thesis (Brno Univ. of Technology, 2012).

    Google Scholar 

  5. M. Kudinov, “Recurrent neural networks for hypotheses re-scoring,” in Proc. Speech and Computer Int. Conf. SPECOM 2015 (Athens, Sept. 20–24, 2015).

    Google Scholar 

  6. S. Hochreiter and J. Schmidhuber, “Bridging long time lags by weight guessing and Long Short-Term Memory,” in Spatiotemporal Models in Biological and Artificial Systems, Ed. by F. Silva, J. Principe, and L. Almeida (IOS Press, 1996).

    Google Scholar 

  7. R. Pascanu, T. Mikolov, and Y. Bengio, On the difficulty of training recurrent neural networks (2012). arXiv:1211.5063

    Google Scholar 

  8. R. Pascanu, C. Gulcehre, K. Cho, and Y. Bengio, “How to construct deep recurrent neural networks,” in Proc. ICLR 2014 (Banff, 2014). arXiv:1312.6026

    Google Scholar 

  9. Y. Bengio, P. Simard, and P. Frasconi, “Learning longterm dependencies with gradient descent is difficult,” IEEE Trans. Neural Networks 5 (2), 157–166 (1994).

    Article  Google Scholar 

  10. T. Mikolov and G. Zweig, “Context dependent recurrent neural network language model,” in Proc. IEEE Spoken Language Technology Workshop (Miami, 2012).

    Google Scholar 

  11. D. Blei, A. Ng, and M. Jordan, “Latent dirichlet allocation,” J. Mach. Learn. Res. 3, 993–1022 (2003).

    MATH  Google Scholar 

  12. J. Bellegarda, “Exploiting latent semantic information in statistical language modeling,” Proc. IEEE. 88 (8), 1279 (2000).

    Article  Google Scholar 

  13. D. Gildea and T. Hoffman, “Topic-based language models using EM,” in Proc. EUROSPEECH (Budapest, 1999).

    Google Scholar 

  14. T. Hofmann, “Probabilistic latent semantic indexing,” in Proc. 22nd Annu. Int. ACM SIGIR Conf. on Research and Development in Information Retrieval (ACM, New York, 1999), pp. 50–57.

    Google Scholar 

  15. K.V.Vorontsov, Probabilistic Thematic Simulation. http://www.machinelearning.ru/wiki/images/2/22/Voron-2013-ptm.pdf. Cited September 29, 2015.

  16. K. Vorontsov and A. Potapenko, “Regularization, robustness, and rarefiety of probabilistic thematic models,” Komp’yut. Issl. Model. 4 (4), 693–706 (2012).

    Google Scholar 

  17. R. Rosenfeld, “A maximum entropy approach to adaptive statistical language modelling,” Comput. Speech Language 10 (3), 187–228 (1996).

    Article  MathSciNet  Google Scholar 

  18. S. Muzychka, A. Romanenko, and I. Piontkovskaja, “Conditional random field for morphological disambiguation in Russian,” in Proc. Conf. Dialog-2014 (Bekasovo, 2014).

    Google Scholar 

  19. A. N. Tikhonov and V. Ya. Arsenin, Methods for Solving Incorrect Problems (Nauka, Moscow, 1986) [in Russian].

    Google Scholar 

  20. K. Vorontsov and A. Potapenko, “Probabilistic thematic models regularization for rising interpretability and determining the number of topics,” in Proc. Conf. Dialog-2014 (Bekasovo, 2014).

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to M. S. Kudinov.

Additional information

This article was translated by the authors.

Mikhail Sergeevich Kudinov. Born 1990. Student in Theoretical and Computational Linguistics, Faculty of Philology, Moscow State University, 2007–2012. Since 2012 graduate student at Dorodnitsyn Computing Centre, Russian Academy of Sciences. Fields of interest: natural language processing, automatic speech recognition, language modeling, machine learning. Author of 6 papers.

Aleksandr Aleksandrovich Romanenko. Born 1991. Student at the Department of Control and Applied Mathematics, Moscow Institute of Physics and Technology, 2008–2014. Awarded Master’s degree in 2014. Since 2015 graduate student at Moscow Institute of Physics and Technology, academic adviser K.V.Vorontsov. Fields of interest: probabilistic topic modeling, machine learning, document analysis. Author of 5 papers.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Kudinov, M.S., Romanenko, A.A. A hybrid language model based on a recurrent neural network and probabilistic topic modeling. Pattern Recognit. Image Anal. 26, 587–592 (2016). https://doi.org/10.1134/S1054661816030123

Download citation

  • Received:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1134/S1054661816030123

Keywords

Navigation