Skip to main content
Top

2018 | OriginalPaper | Chapter

9. Neural Language Models

Author : Sandro Skansi

Published in: Introduction to Deep Learning

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

This chapter revisits language processing, this time equipped with deep learning. Recurrent neural networks and autoencoders are needed for this chapter, but the exposition is clear and uses them mainly in a conceptual rather than computational sense. The idea of word embeddings is explored and the main deep learning method for representing text, the neural word embedding is described with the famous Word2vec algorithm, in both the Skip-gram and CBOW variant. A CBOW Word2vec architecture is explored in detail and presented in Python code. This code presupposes a text as a list, but this code was written and explained in the previous chapters, and PCA is used in the code to reduce the dimensionality of the vectors to enable an easy display. The chapter concludes with word analogies and simple calculations that can be done and form the basis of analogical reasoning, a simple reasoning calculus that is neural all the way with no symbolic manipulation used.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Footnotes
1
If the context were 2, it would take 4 words, two before the main word and two after.
 
2
If we were to save and load from a H5 file, we would be saving ans loading all the weights in a new network of the same configuration, possibly fine-tuning them and then taking out just the weight matrix with the same code we used here.
 
3
More precisely: to transform the matrix into a decorrelated matrix whose columns are arranged in descending variance and then keep the first two columns.
 
Literature
2.
go back to reference V.I. Levenshtein, Binary codes capable of correcting deletions, insertions, and reversals. Sov. Phys. Dokl. 10(8), 707–710 (1966)MathSciNet V.I. Levenshtein, Binary codes capable of correcting deletions, insertions, and reversals. Sov. Phys. Dokl. 10(8), 707–710 (1966)MathSciNet
3.
go back to reference M.A. Jaro, Advances in record linkage methodology as applied to the 1985 census of tampa florida. J. Am. Stat. Assoc. 84(406), 414–420 (1989)CrossRef M.A. Jaro, Advances in record linkage methodology as applied to the 1985 census of tampa florida. J. Am. Stat. Assoc. 84(406), 414–420 (1989)CrossRef
4.
go back to reference W.E. Winkler, String comparator metrics and enhanced decision rules in the fellegi-sunter model of record linkage, in Proceedings of the Section on Survey Research Methods (American Statistical Association, 1990), pp. 354–359 W.E. Winkler, String comparator metrics and enhanced decision rules in the fellegi-sunter model of record linkage, in Proceedings of the Section on Survey Research Methods (American Statistical Association, 1990), pp. 354–359
5.
go back to reference A. Singhal, Modern information retrieval: a brief overview. Bull. IEEE Comput. Soc. Tech. Comm. Data Eng. 24(4), 35–43 (2001) A. Singhal, Modern information retrieval: a brief overview. Bull. IEEE Comput. Soc. Tech. Comm. Data Eng. 24(4), 35–43 (2001)
6.
go back to reference T. Mikolov, T. Chen, G. Corrado, J. Dean, Efficient estimation of word representations in vector space, in ICLR Workshop (2013), arXiv:1301.3781 T. Mikolov, T. Chen, G. Corrado, J. Dean, Efficient estimation of word representations in vector space, in ICLR Workshop (2013), arXiv:​1301.​3781
7.
8.
go back to reference J.R. Firth, A synopsis of linguistic theory 1930–1955, in Studies in Linguistic Analysis (Philological Society, 1957), pp. 1–32 J.R. Firth, A synopsis of linguistic theory 1930–1955, in Studies in Linguistic Analysis (Philological Society, 1957), pp. 1–32
9.
go back to reference L. Wittgenstein, Philosophical Investigations (MacMillan Publishing Company, London, 1953)MATH L. Wittgenstein, Philosophical Investigations (MacMillan Publishing Company, London, 1953)MATH
10.
go back to reference H. Moravec, Mind Children: The Future of Robot and Human Intelligence (Harvard University Press, Cambridge, 1988) H. Moravec, Mind Children: The Future of Robot and Human Intelligence (Harvard University Press, Cambridge, 1988)
Metadata
Title
Neural Language Models
Author
Sandro Skansi
Copyright Year
2018
DOI
https://doi.org/10.1007/978-3-319-73004-2_9

Premium Partner