Skip to main content
Erschienen in: Neural Computing and Applications 36/2023

18.06.2023 | S.I.: Evolutionary Computation based Methods and Applications for Data Processing

ECBTNet: English-Foreign Chinese intelligent translation via multi-subspace attention and hyperbolic tangent LSTM

verfasst von: Jing Yang

Erschienen in: Neural Computing and Applications | Ausgabe 36/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

The translation and sharing of languages around the world has become a necessary precondition for the movement of people. Teaching Chinese as a foreign language (TCFL) undertakes international function of spreading national culture. How to translate Chinese as a foreign language into English has become an important task. Machine translation has moved beyond the realm of theory to practical use as a result of advancements in computing. Deep learning is a prominent and relatively young subfield of machine learning that has shown promising results in a variety of fields. This paper aims to develop a TCFL-oriented English-Chinese neural machine translation model. First, this paper proposes a hyperbolic tangent long short-term memory network (HTLSTM). This will integrate future information and historical information to extract more sufficient contextual semantic information. Secondly, this paper proposes a multi-subspace attention mechanism. This integrates multiple attention calculation functions in the multi-subspace attention mechanism (MSATT). Thirdly, this paper combines HTLSTM with MSATT to construct an English-Chinese bilingual neural translation model called ECBTNet. The multi-subspace attention maps hidden state of hyperbolic tangent long-term short-term memory network to multiple subspaces. This then uses multiple attention calculation functions in the multi-attention mechanism when calculating the attention score. By applying different attention calculation functions in different subspaces to extract omni-directional context information features, accurate attention calculation results can be obtained. Finally, a systematic experiment is carried out, and the experimental data verify the feasibility of applying ECBTNet to the field of English-Chinese translation in TCFL.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Rivera-Trigueros I (2022) Machine translation systems and quality assessment: a systematic review[J]. Lang Resour Eval 56(2):593–619CrossRef Rivera-Trigueros I (2022) Machine translation systems and quality assessment: a systematic review[J]. Lang Resour Eval 56(2):593–619CrossRef
2.
Zurück zum Zitat Klimova B, Pikhart M, Benites AD et al (2023) Neural machine translation in foreign language teaching and learning: a systematic review[J]. Educ Inf Technol 28(1):663–682CrossRef Klimova B, Pikhart M, Benites AD et al (2023) Neural machine translation in foreign language teaching and learning: a systematic review[J]. Educ Inf Technol 28(1):663–682CrossRef
3.
Zurück zum Zitat Ranathunga S, Lee ESA, Prifti Skenduli M et al (2023) Neural machine translation for low-resource languages: a survey[J]. ACM Comput Surv 55(11):1–37CrossRef Ranathunga S, Lee ESA, Prifti Skenduli M et al (2023) Neural machine translation for low-resource languages: a survey[J]. ACM Comput Surv 55(11):1–37CrossRef
4.
Zurück zum Zitat Lee SM (2023) The effectiveness of machine translation in foreign language education: a systematic review and meta-analysis[J]. Comput Assist Lang Learn 36(1–2):103–125CrossRef Lee SM (2023) The effectiveness of machine translation in foreign language education: a systematic review and meta-analysis[J]. Comput Assist Lang Learn 36(1–2):103–125CrossRef
5.
Zurück zum Zitat Guerberof-Arenas A, Toral A (2022) Creativity in translation: machine translation as a constraint for literary texts[J]. Transl Spaces 11(2):184–212CrossRef Guerberof-Arenas A, Toral A (2022) Creativity in translation: machine translation as a constraint for literary texts[J]. Transl Spaces 11(2):184–212CrossRef
7.
Zurück zum Zitat Ryu J, Kim Y, Park S, et al. (2022) Exploring foreign language students’ perceptions of the guided use of machine translation (GUMT) model for Korean writing[J]. L2 J. 14(1) Ryu J, Kim Y, Park S, et al. (2022) Exploring foreign language students’ perceptions of the guided use of machine translation (GUMT) model for Korean writing[J]. L2 J. 14(1)
8.
Zurück zum Zitat Mondal SK, Zhang H, Kabir HMD et al (2023) Machine translation and its evaluation: a study[J]. Artif Intell Rev 1:1–90 Mondal SK, Zhang H, Kabir HMD et al (2023) Machine translation and its evaluation: a study[J]. Artif Intell Rev 1:1–90
9.
Zurück zum Zitat Pei J, Zhong K, Yu Z, et al. (2022) Scene graph semantic inference for image and text matching[J]. Transactions on Asian and Low-Resource Language Information Processing, 1 Pei J, Zhong K, Yu Z, et al. (2022) Scene graph semantic inference for image and text matching[J]. Transactions on Asian and Low-Resource Language Information Processing, 1
10.
Zurück zum Zitat Saunders D (2022) Domain adaptation and multi-domain adaptation for neural machine translation: a survey[J]. J Artif Intell Res 75:351–424MathSciNetCrossRefMATH Saunders D (2022) Domain adaptation and multi-domain adaptation for neural machine translation: a survey[J]. J Artif Intell Res 75:351–424MathSciNetCrossRefMATH
11.
Zurück zum Zitat Samant RM, Bachute MR, Gite S et al (2022) Framework for deep learning-based language models using multi-task learning in natural language understanding: a systematic literature review and future directions[J]. IEEE Access 10:17078–17097CrossRef Samant RM, Bachute MR, Gite S et al (2022) Framework for deep learning-based language models using multi-task learning in natural language understanding: a systematic literature review and future directions[J]. IEEE Access 10:17078–17097CrossRef
12.
Zurück zum Zitat Dabre R, Chu C, Kunchukuttan A (2020) A survey of multilingual neural machine translation[J]. ACM Comput Surv (CSUR) 53(5):1–38CrossRef Dabre R, Chu C, Kunchukuttan A (2020) A survey of multilingual neural machine translation[J]. ACM Comput Surv (CSUR) 53(5):1–38CrossRef
13.
Zurück zum Zitat Andrabi SAB, Wahid A (2022) Machine translation system using deep learning for English to Urdu[J]. Comput Intell Neurosci Andrabi SAB, Wahid A (2022) Machine translation system using deep learning for English to Urdu[J]. Comput Intell Neurosci
14.
Zurück zum Zitat Al-Sayed MM (2022) Workload time series cumulative prediction mechanism for cloud resources using neural machine translation technique[J]. J Grid Comput 20(2):16CrossRef Al-Sayed MM (2022) Workload time series cumulative prediction mechanism for cloud resources using neural machine translation technique[J]. J Grid Comput 20(2):16CrossRef
15.
Zurück zum Zitat Nguyen PT, Di Rocco J, Rubei R et al (2022) DeepLib: Machine translation techniques to recommend upgrades for third-party libraries[J]. Expert Syst Appl 202:117267CrossRef Nguyen PT, Di Rocco J, Rubei R et al (2022) DeepLib: Machine translation techniques to recommend upgrades for third-party libraries[J]. Expert Syst Appl 202:117267CrossRef
16.
Zurück zum Zitat Bensalah N, Ayad H, Adib A, et al. (2022) CRAN: an hybrid CNN-RNN attention-based model for Arabic machine translation[C]. Networking, Intelligent Systems and Security: Proceedings of NISS 2021. Springer Singapore, 87–102 Bensalah N, Ayad H, Adib A, et al. (2022) CRAN: an hybrid CNN-RNN attention-based model for Arabic machine translation[C]. Networking, Intelligent Systems and Security: Proceedings of NISS 2021. Springer Singapore, 87–102
17.
Zurück zum Zitat Chiche A, Yitagesu B (2022) Part of speech tagging: a systematic review of deep learning and machine learning approaches[J]. J Big Data 9(1):1–25CrossRef Chiche A, Yitagesu B (2022) Part of speech tagging: a systematic review of deep learning and machine learning approaches[J]. J Big Data 9(1):1–25CrossRef
18.
Zurück zum Zitat Fan A, Bhosale S, Schwenk H et al (2021) Beyond english-centric multilingual machine translation[J]. J Mach Learn Res 22(1):4839–4886MathSciNetMATH Fan A, Bhosale S, Schwenk H et al (2021) Beyond english-centric multilingual machine translation[J]. J Mach Learn Res 22(1):4839–4886MathSciNetMATH
19.
Zurück zum Zitat Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks[J]. Adv Neural Inf Process Syst 27:3104–3112 Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks[J]. Adv Neural Inf Process Syst 27:3104–3112
20.
Zurück zum Zitat Cho K, van Merriënboer B, Gulcehre C, et al. (2014) Learning Phrase Representations Using RNN Encoder–Decoder for Statistical Machine Translation[C]. Conference on Empirical Methods in Natural Language Processing, 1724–1734 Cho K, van Merriënboer B, Gulcehre C, et al. (2014) Learning Phrase Representations Using RNN Encoder–Decoder for Statistical Machine Translation[C]. Conference on Empirical Methods in Natural Language Processing, 1724–1734
21.
Zurück zum Zitat Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate[J]. arXiv preprint arXiv:1409.0473 Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate[J]. arXiv preprint arXiv:​1409.​0473
22.
Zurück zum Zitat Luong M T, Pham H, Manning C D (2015) Effective approaches to attention-based neural machine translation[C]. Conference on Empirical Methods in Natural Language Processing 1412–1421 Luong M T, Pham H, Manning C D (2015) Effective approaches to attention-based neural machine translation[C]. Conference on Empirical Methods in Natural Language Processing 1412–1421
23.
Zurück zum Zitat Jean S, Cho K, Memisevic R., Bengio, Y (2015) On using very large target vocabulary for neural machine translation[C]. Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing 1–10 Jean S, Cho K, Memisevic R., Bengio, Y (2015) On using very large target vocabulary for neural machine translation[C]. Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing 1–10
24.
Zurück zum Zitat Junczys-Dowmunt M, Dwojak T, Hoang H (2016) Is neural machine translation ready for deployment[J]. A case study on, 30 Junczys-Dowmunt M, Dwojak T, Hoang H (2016) Is neural machine translation ready for deployment[J]. A case study on, 30
25.
Zurück zum Zitat Gehring J, Auli M, Grangier D, et al. (2017) Convolutional sequence to sequence learning[C]. International conference on machine learning 1243–1252 Gehring J, Auli M, Grangier D, et al. (2017) Convolutional sequence to sequence learning[C]. International conference on machine learning 1243–1252
26.
Zurück zum Zitat Sennrich R, Haddow B, Birch A (2016) Neural machine translation of rare words with subword units[C]. Annual Meeting of the Association for Computational Linguistics 1715–1725 Sennrich R, Haddow B, Birch A (2016) Neural machine translation of rare words with subword units[C]. Annual Meeting of the Association for Computational Linguistics 1715–1725
27.
Zurück zum Zitat Vaswani A, Shazeer N, Parmar N, et al. (2017) Attention is All You Need[C]. International Conference on Neural Information Processing Systems, 6000–6010. Vaswani A, Shazeer N, Parmar N, et al. (2017) Attention is All You Need[C]. International Conference on Neural Information Processing Systems, 6000–6010.
28.
Zurück zum Zitat Hassan H, Aue A, Chen C, et al. (2018) Achieving human parity on automatic chinese to english news translation[J]. arXiv preprint arXiv:1803.05567 Hassan H, Aue A, Chen C, et al. (2018) Achieving human parity on automatic chinese to english news translation[J]. arXiv preprint arXiv:​1803.​05567
30.
Zurück zum Zitat Dai Z, Yang Z, Yang Y, et al. (2019) Transformer-XL: attentive language models beyond a fixed-length context[C]. Annual Meeting of the Association for Computational Linguistics 2978–2988 Dai Z, Yang Z, Yang Y, et al. (2019) Transformer-XL: attentive language models beyond a fixed-length context[C]. Annual Meeting of the Association for Computational Linguistics 2978–2988
31.
Zurück zum Zitat Wang Q, Li B, Xiao T, et al. (2019) Learning deep transformer models for machine translation[C]. Annual Meeting of the Association for Computational Linguistics., 1810–1822 Wang Q, Li B, Xiao T, et al. (2019) Learning deep transformer models for machine translation[C]. Annual Meeting of the Association for Computational Linguistics., 1810–1822
32.
Zurück zum Zitat Dedes K, Utama ABP, Wibawa AP et al. (2022) Neural machine translation of Spanish-English food recipes using LSTM[J]. JOIV: Int J Informat Visual 6(2):290–297 Dedes K, Utama ABP, Wibawa AP et al. (2022) Neural machine translation of Spanish-English food recipes using LSTM[J]. JOIV: Int J Informat Visual 6(2):290–297
33.
Zurück zum Zitat Xiao Q, Chang X, Zhang X et al (2020) Multi-information spatial–temporal LSTM fusion continuous sign language neural machine translation[J]. IEEE Access 8:216718–216728CrossRef Xiao Q, Chang X, Zhang X et al (2020) Multi-information spatial–temporal LSTM fusion continuous sign language neural machine translation[J]. IEEE Access 8:216718–216728CrossRef
34.
Zurück zum Zitat Sartipi A, Dehghan M, Fatemi A (2023) An evaluation of persian-english machine translation datasets with transformers[J]. arXiv preprint arXiv:2302.00321 Sartipi A, Dehghan M, Fatemi A (2023) An evaluation of persian-english machine translation datasets with transformers[J]. arXiv preprint arXiv:​2302.​00321
Metadaten
Titel
ECBTNet: English-Foreign Chinese intelligent translation via multi-subspace attention and hyperbolic tangent LSTM
verfasst von
Jing Yang
Publikationsdatum
18.06.2023
Verlag
Springer London
Erschienen in
Neural Computing and Applications / Ausgabe 36/2023
Print ISSN: 0941-0643
Elektronische ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-023-08624-8

Weitere Artikel der Ausgabe 36/2023

Neural Computing and Applications 36/2023 Zur Ausgabe

S.I.: Evolutionary Computation based Methods and Applications for Data Processing

Localization algorithm for anisotropic wireless sensor networks based on the adaptive chaotic slime mold algorithm

S.I.: Evolutionary Computation-based Methods and Applications for Data Processing

Research on sheep face recognition algorithm based on improved AlexNet model

S.I.: Evolutionary Computation based Methods and Applications for Data Processing

Ontology construction and mapping of multi-source heterogeneous data based on hybrid neural network and autoencoder

S.I. : Evolutionary Computation based Methods and Applications for Data Processing

Modeling the gaze point distribution to assist eye-based target selection in head-mounted displays

S.I.: Evolutionary Computation based Methods and Applications for Data Processing

A graph-based collaborative filtering algorithm combining implicit user preference and explicit time-related feedback

S.I.: Evolutionary Computation based Methods and Applications for Data Processing

Dynamic model averaging-based procurement optimization of prefabricated components

Premium Partner