Skip to main content
Top

2019 | OriginalPaper | Chapter

Improved Neural Machine Translation with POS-Tagging Through Joint Decoding

Authors : Xiaocheng Feng, Zhangyin Feng, Wanlong Zhao, Nan Zou, Bing Qin, Ting Liu

Published in: Artificial Intelligence for Communications and Networks

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

In this paper, we improve the performance of neural machine translation (NMT) with shallow syntax (e.g., POS tag) of target language, which has better accuracy and latency than deep syntax such as dependency parsing. We present three NMT decoding models (independent decoder, gates shared decoder and fully shared decoder) to jointly predict target word and POS tag sequences. Experiments on Chinese-English and German-English translation tasks show that the fully shared decoder can acquire the best performance, which increases the BLEU score by 1.4 and 2.25 points respectively compared with the attention-based NMT model.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Footnotes
1
Six combinations (shared gates / independent gates): {[input / forget, output], [input, forget / output], [input,output / forget], [forget / input, output], [output / input, forget], [forget, output / input]}.
 
2
The codes are implemented with Pytorch, which we plan to release to the community.
 
3
The corpora includes LDC2002E18, LDC2003E07, LDC2003E14, the Hansards portion of LDC2004T08, and LDC2005T06.
 
5
The value of kappa is 0.65 in 1–5 scale on two dimensions.
 
Literature
1.
go back to reference LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436 (2015)CrossRef LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436 (2015)CrossRef
2.
go back to reference Bentivogli, L., Bisazza, A., Cettolo, M., et al.: Neural versus phrase-based machine translation quality: a case study (2016) Bentivogli, L., Bisazza, A., Cettolo, M., et al.: Neural versus phrase-based machine translation quality: a case study (2016)
3.
go back to reference Eriguchi, A., Tsuruoka, Y., Cho, K.: Learning to parse and translate improves neural machine translation (2017) Eriguchi, A., Tsuruoka, Y., Cho, K.: Learning to parse and translate improves neural machine translation (2017)
4.
go back to reference Hashimoto, K., Tsuruoka, Y.: Neural machine translation with source-side latent graph parsing. In: Proceedings of the Conference on Machine Translation, WMT 2017, pp. 125–135. Association for Computational Linguistics (2017) Hashimoto, K., Tsuruoka, Y.: Neural machine translation with source-side latent graph parsing. In: Proceedings of the Conference on Machine Translation, WMT 2017, pp. 125–135. Association for Computational Linguistics (2017)
5.
go back to reference Nadejde, M., Reddy, S., Sennrich, R., et al.: Predicting target language CCG Supertags improves neural machine translation. In: Proceedings of the Conference on Machine Translation, WMT 2017, vol. 1, pp. 68–79. Association for Computational Linguistics (2017) Nadejde, M., Reddy, S., Sennrich, R., et al.: Predicting target language CCG Supertags improves neural machine translation. In: Proceedings of the Conference on Machine Translation, WMT 2017, vol. 1, pp. 68–79. Association for Computational Linguistics (2017)
6.
go back to reference Luong, M.-T., Le, Q.V., Sutskever, I., Vinyals, O., Kaiser, L.: Multi-task sequence to sequence learning. In: Proceedings of the Conference on Machine Translation, WMT 2016. ICLR (2016) Luong, M.-T., Le, Q.V., Sutskever, I., Vinyals, O., Kaiser, L.: Multi-task sequence to sequence learning. In: Proceedings of the Conference on Machine Translation, WMT 2016. ICLR (2016)
7.
go back to reference Niehues, J., Cho E.: Exploiting linguistic resources for neural machine translation using multi-task learning. In: Proceedings of the Conference on Machine Translation, WMT 2017, vol. 1, pp. 80–89. Association for Computational Linguistics (2017) Niehues, J., Cho E.: Exploiting linguistic resources for neural machine translation using multi-task learning. In: Proceedings of the Conference on Machine Translation, WMT 2017, vol. 1, pp. 80–89. Association for Computational Linguistics (2017)
8.
go back to reference Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRef Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRef
9.
go back to reference Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate (2014) Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate (2014)
10.
go back to reference Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: 27th Annual Conference on Neural Information Processing Systems 2013, pp. 3111–3119 (2013) Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: 27th Annual Conference on Neural Information Processing Systems 2013, pp. 3111–3119 (2013)
11.
go back to reference Manning, C., Surdeanu, M., Bauer, J., Finkel, J., Bethard, S., McClosky, D.: The Stanford CoreNLP natural language processing toolkit. In: Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pp. 55–60 (2014) Manning, C., Surdeanu, M., Bauer, J., Finkel, J., Bethard, S., McClosky, D.: The Stanford CoreNLP natural language processing toolkit. In: Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pp. 55–60 (2014)
12.
go back to reference Ranzato, M., Chopra, S., Auli, M., et al.: Sequence level training with recurrent neural networks. Comput. Sci. (2015) Ranzato, M., Chopra, S., Auli, M., et al.: Sequence level training with recurrent neural networks. Comput. Sci. (2015)
13.
go back to reference Zhou, H., Tu, Z., Huang, S., et al.: Chunk-based bi-scale decoder for neural machine translation. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Short Papers), pp. 580–586. Association for Computational Linguistics (2017) Zhou, H., Tu, Z., Huang, S., et al.: Chunk-based bi-scale decoder for neural machine translation. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Short Papers), pp. 580–586. Association for Computational Linguistics (2017)
Metadata
Title
Improved Neural Machine Translation with POS-Tagging Through Joint Decoding
Authors
Xiaocheng Feng
Zhangyin Feng
Wanlong Zhao
Nan Zou
Bing Qin
Ting Liu
Copyright Year
2019
DOI
https://doi.org/10.1007/978-3-030-22968-9_14

Premium Partner