Skip to main content

2019 | OriginalPaper | Buchkapitel

Improved Neural Machine Translation with POS-Tagging Through Joint Decoding

verfasst von : Xiaocheng Feng, Zhangyin Feng, Wanlong Zhao, Nan Zou, Bing Qin, Ting Liu

Erschienen in: Artificial Intelligence for Communications and Networks

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

In this paper, we improve the performance of neural machine translation (NMT) with shallow syntax (e.g., POS tag) of target language, which has better accuracy and latency than deep syntax such as dependency parsing. We present three NMT decoding models (independent decoder, gates shared decoder and fully shared decoder) to jointly predict target word and POS tag sequences. Experiments on Chinese-English and German-English translation tasks show that the fully shared decoder can acquire the best performance, which increases the BLEU score by 1.4 and 2.25 points respectively compared with the attention-based NMT model.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
Six combinations (shared gates / independent gates): {[input / forget, output], [input, forget / output], [input,output / forget], [forget / input, output], [output / input, forget], [forget, output / input]}.
 
2
The codes are implemented with Pytorch, which we plan to release to the community.
 
3
The corpora includes LDC2002E18, LDC2003E07, LDC2003E14, the Hansards portion of LDC2004T08, and LDC2005T06.
 
5
The value of kappa is 0.65 in 1–5 scale on two dimensions.
 
Literatur
1.
Zurück zum Zitat LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436 (2015)CrossRef LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436 (2015)CrossRef
2.
Zurück zum Zitat Bentivogli, L., Bisazza, A., Cettolo, M., et al.: Neural versus phrase-based machine translation quality: a case study (2016) Bentivogli, L., Bisazza, A., Cettolo, M., et al.: Neural versus phrase-based machine translation quality: a case study (2016)
3.
Zurück zum Zitat Eriguchi, A., Tsuruoka, Y., Cho, K.: Learning to parse and translate improves neural machine translation (2017) Eriguchi, A., Tsuruoka, Y., Cho, K.: Learning to parse and translate improves neural machine translation (2017)
4.
Zurück zum Zitat Hashimoto, K., Tsuruoka, Y.: Neural machine translation with source-side latent graph parsing. In: Proceedings of the Conference on Machine Translation, WMT 2017, pp. 125–135. Association for Computational Linguistics (2017) Hashimoto, K., Tsuruoka, Y.: Neural machine translation with source-side latent graph parsing. In: Proceedings of the Conference on Machine Translation, WMT 2017, pp. 125–135. Association for Computational Linguistics (2017)
5.
Zurück zum Zitat Nadejde, M., Reddy, S., Sennrich, R., et al.: Predicting target language CCG Supertags improves neural machine translation. In: Proceedings of the Conference on Machine Translation, WMT 2017, vol. 1, pp. 68–79. Association for Computational Linguistics (2017) Nadejde, M., Reddy, S., Sennrich, R., et al.: Predicting target language CCG Supertags improves neural machine translation. In: Proceedings of the Conference on Machine Translation, WMT 2017, vol. 1, pp. 68–79. Association for Computational Linguistics (2017)
6.
Zurück zum Zitat Luong, M.-T., Le, Q.V., Sutskever, I., Vinyals, O., Kaiser, L.: Multi-task sequence to sequence learning. In: Proceedings of the Conference on Machine Translation, WMT 2016. ICLR (2016) Luong, M.-T., Le, Q.V., Sutskever, I., Vinyals, O., Kaiser, L.: Multi-task sequence to sequence learning. In: Proceedings of the Conference on Machine Translation, WMT 2016. ICLR (2016)
7.
Zurück zum Zitat Niehues, J., Cho E.: Exploiting linguistic resources for neural machine translation using multi-task learning. In: Proceedings of the Conference on Machine Translation, WMT 2017, vol. 1, pp. 80–89. Association for Computational Linguistics (2017) Niehues, J., Cho E.: Exploiting linguistic resources for neural machine translation using multi-task learning. In: Proceedings of the Conference on Machine Translation, WMT 2017, vol. 1, pp. 80–89. Association for Computational Linguistics (2017)
8.
Zurück zum Zitat Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRef Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRef
9.
Zurück zum Zitat Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate (2014) Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate (2014)
10.
Zurück zum Zitat Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: 27th Annual Conference on Neural Information Processing Systems 2013, pp. 3111–3119 (2013) Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: 27th Annual Conference on Neural Information Processing Systems 2013, pp. 3111–3119 (2013)
11.
Zurück zum Zitat Manning, C., Surdeanu, M., Bauer, J., Finkel, J., Bethard, S., McClosky, D.: The Stanford CoreNLP natural language processing toolkit. In: Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pp. 55–60 (2014) Manning, C., Surdeanu, M., Bauer, J., Finkel, J., Bethard, S., McClosky, D.: The Stanford CoreNLP natural language processing toolkit. In: Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pp. 55–60 (2014)
12.
Zurück zum Zitat Ranzato, M., Chopra, S., Auli, M., et al.: Sequence level training with recurrent neural networks. Comput. Sci. (2015) Ranzato, M., Chopra, S., Auli, M., et al.: Sequence level training with recurrent neural networks. Comput. Sci. (2015)
13.
Zurück zum Zitat Zhou, H., Tu, Z., Huang, S., et al.: Chunk-based bi-scale decoder for neural machine translation. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Short Papers), pp. 580–586. Association for Computational Linguistics (2017) Zhou, H., Tu, Z., Huang, S., et al.: Chunk-based bi-scale decoder for neural machine translation. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Short Papers), pp. 580–586. Association for Computational Linguistics (2017)
Metadaten
Titel
Improved Neural Machine Translation with POS-Tagging Through Joint Decoding
verfasst von
Xiaocheng Feng
Zhangyin Feng
Wanlong Zhao
Nan Zou
Bing Qin
Ting Liu
Copyright-Jahr
2019
DOI
https://doi.org/10.1007/978-3-030-22968-9_14