Skip to main content
Top

2024 | OriginalPaper | Chapter

MTLAN: Multi-Task Learning and Auxiliary Network for Enhanced Sentence Embedding

Authors : Gang Liu, Tongli Wang, Wenli Yang, Zhizheng Yan, Kai Zhan

Published in: Neural Information Processing

Publisher: Springer Nature Singapore

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

The objective of cross-lingual sentence embedding learning is to map sentences into a shared representation space, where semantically similar sentence representations are closer together, while distinct sentence representations exhibit clear differentiation. This paper proposes a novel sentence embedding model called MTLAN, which incorporates multi-task learning and auxiliary networks. The model utilizes the LaBSE model for extracting sentence features and undergoes joint training on tasks related to sentence semantic representation and distance measurement. Furthermore, an auxiliary network is employed to enhance the contextual expression of words within sentences. To address the issue of limited resources for low-resource languages, we construct a pseudo-corpus dataset using a multilingual dictionary for unsupervised learning. We conduct experiments on multiple publicly available datasets, including STS and SICK, to evaluate both monolingual sentence similarity and cross-lingual semantic similarity. The empirical results demonstrate the significant superiority of our proposed model over state-of-the-art methods.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Pham, H., Luong, M.T., Manning, C.D.: Learning distributed representations for multilingual text sequences. In: Proceedings of the 1st Workshop on Vector Space Modeling for Natural Language Processing, pp. 88–94 (2015) Pham, H., Luong, M.T., Manning, C.D.: Learning distributed representations for multilingual text sequences. In: Proceedings of the 1st Workshop on Vector Space Modeling for Natural Language Processing, pp. 88–94 (2015)
3.
go back to reference Conneau, A., Lample, G.: Cross-lingual language model pretraining 32 (2019) Conneau, A., Lample, G.: Cross-lingual language model pretraining 32 (2019)
4.
go back to reference Conneau, A., et al.: Unsupervised cross-lingual representation learning at scale, pp. 8440–8451 (2020) Conneau, A., et al.: Unsupervised cross-lingual representation learning at scale, pp. 8440–8451 (2020)
5.
go back to reference Liu, J., et al.: RankCSE: Unsupervised sentence representations learning via learning to rank (2023) Liu, J., et al.: RankCSE: Unsupervised sentence representations learning via learning to rank (2023)
6.
go back to reference Nie, Z., Zhang, R., Mao, Y.: On the inadequacy of optimizing alignment and uniformity in contrastive learning of sentence representations. In: The Eleventh International Conference on Learning Representations (2023) Nie, Z., Zhang, R., Mao, Y.: On the inadequacy of optimizing alignment and uniformity in contrastive learning of sentence representations. In: The Eleventh International Conference on Learning Representations (2023)
7.
go back to reference Agirre, E., Cer, D., Diab, M., Gonzalez-Agirre, A.: Semeval-2012 task 6: a pilot on semantic textual similarity. In: *SEM 2012: The First Joint Conference on Lexical and Computational Semantics-Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012), pp. 385–393 (2012) Agirre, E., Cer, D., Diab, M., Gonzalez-Agirre, A.: Semeval-2012 task 6: a pilot on semantic textual similarity. In: *SEM 2012: The First Joint Conference on Lexical and Computational Semantics-Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012), pp. 385–393 (2012)
8.
go back to reference Cer, D., Diab, M., Agirre, E., Lopez-Gazpio, I., Specia, L.: Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation. arXiv preprint arXiv:1708.00055 (2017) Cer, D., Diab, M., Agirre, E., Lopez-Gazpio, I., Specia, L.: Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation. arXiv preprint arXiv:​1708.​00055 (2017)
9.
go back to reference Marelli, M., Menini, S., Baroni, M., Bentivogli, L., Bernardi, R., Zamparelli, R., et al.: A sick cure for the evaluation of compositional distributional semantic models. In: LREC, pp. 216–223. Reykjavik (2014) Marelli, M., Menini, S., Baroni, M., Bentivogli, L., Bernardi, R., Zamparelli, R., et al.: A sick cure for the evaluation of compositional distributional semantic models. In: LREC, pp. 216–223. Reykjavik (2014)
10.
go back to reference Artetxe, M., Schwenk, H.: Massively multilingual sentence embeddings for zero-shot cross-lingual transfer and beyond. Trans. Assoc. Comput. Linguist. 7, 597–610 (2019)CrossRef Artetxe, M., Schwenk, H.: Massively multilingual sentence embeddings for zero-shot cross-lingual transfer and beyond. Trans. Assoc. Comput. Linguist. 7, 597–610 (2019)CrossRef
11.
go back to reference Logeswaran, L., Lee, H.: An efficient framework for learning sentence representations (2018) Logeswaran, L., Lee, H.: An efficient framework for learning sentence representations (2018)
12.
go back to reference Conneau, A., Kiela, D., Schwenk, H., Barrault, L., Bordes, A.: Supervised learning of universal sentence representations from natural language inference data, pp. 670–680 (2017) Conneau, A., Kiela, D., Schwenk, H., Barrault, L., Bordes, A.: Supervised learning of universal sentence representations from natural language inference data, pp. 670–680 (2017)
13.
go back to reference Feng, F., Yang, Y., Cer, D., Arivazhagan, N., Wang, W.: Language-agnostic BERT sentence embedding, pp. 878–891, May 2022 Feng, F., Yang, Y., Cer, D., Arivazhagan, N., Wang, W.: Language-agnostic BERT sentence embedding, pp. 878–891, May 2022
14.
go back to reference Zhang, Y., Yang, Q.: A survey on multi-task learning. IEEE Trans. Knowl. Data Eng. 34(12), 5586–5609 (2021)CrossRef Zhang, Y., Yang, Q.: A survey on multi-task learning. IEEE Trans. Knowl. Data Eng. 34(12), 5586–5609 (2021)CrossRef
15.
go back to reference Lample, G., Conneau, A., Denoyer, L., Ranzato, M.: Unsupervised machine translation using monolingual corpora only (2018) Lample, G., Conneau, A., Denoyer, L., Ranzato, M.: Unsupervised machine translation using monolingual corpora only (2018)
16.
go back to reference Conneau, A., et al.: XNLI: evaluating cross-lingual sentence representations. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 2475–2485 (2018) Conneau, A., et al.: XNLI: evaluating cross-lingual sentence representations. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 2475–2485 (2018)
17.
go back to reference Agirre, E., et al.: Semeval-2014 task 10: multilingual semantic textual similarity. In: Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014), pp. 81–91 (2014) Agirre, E., et al.: Semeval-2014 task 10: multilingual semantic textual similarity. In: Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014), pp. 81–91 (2014)
18.
go back to reference Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding, pp. 4171–4186, June 2019 Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding, pp. 4171–4186, June 2019
19.
go back to reference Yan, Y., Li, R., Wang, S., Zhang, F., Wu, W., Xu, W.: ConSERT: a contrastive framework for self-supervised sentence representation transfer, pp. 5065–5075, August 2021 Yan, Y., Li, R., Wang, S., Zhang, F., Wu, W., Xu, W.: ConSERT: a contrastive framework for self-supervised sentence representation transfer, pp. 5065–5075, August 2021
20.
go back to reference Gao, T., Yao, X., Chen, D.: SimCSE: simple contrastive learning of sentence embeddings, pp. 6894–6910, November 2021 Gao, T., Yao, X., Chen, D.: SimCSE: simple contrastive learning of sentence embeddings, pp. 6894–6910, November 2021
21.
go back to reference Wu, X., Gao, C., Lin, Z., Han, J., Wang, Z., Hu, S.: InfoCSE: information-aggregated contrastive learning of sentence embeddings, pp. 3060–3070, December 2022 Wu, X., Gao, C., Lin, Z., Han, J., Wang, Z., Hu, S.: InfoCSE: information-aggregated contrastive learning of sentence embeddings, pp. 3060–3070, December 2022
22.
go back to reference Jiang, T., et al.: PromptBERT: improving BERT sentence embeddings with prompts, pp. 8826–8837, December 2022 Jiang, T., et al.: PromptBERT: improving BERT sentence embeddings with prompts, pp. 8826–8837, December 2022
23.
go back to reference Seonwoo, Y., et al.: Ranking-enhanced unsupervised sentence representation learning (2023) Seonwoo, Y., et al.: Ranking-enhanced unsupervised sentence representation learning (2023)
24.
go back to reference Goswami, K., Dutta, S., Assem, H., Fransen, T., McCrae, J.P.: Cross-lingual sentence embedding using multi-task learning. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 9099–9113. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic, November 2021 Goswami, K., Dutta, S., Assem, H., Fransen, T., McCrae, J.P.: Cross-lingual sentence embedding using multi-task learning. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 9099–9113. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic, November 2021
Metadata
Title
MTLAN: Multi-Task Learning and Auxiliary Network for Enhanced Sentence Embedding
Authors
Gang Liu
Tongli Wang
Wenli Yang
Zhizheng Yan
Kai Zhan
Copyright Year
2024
Publisher
Springer Nature Singapore
DOI
https://doi.org/10.1007/978-981-99-8067-3_2

Premium Partner