Skip to main content
Top

22-08-2024

Scrutinizing Label: Contrastive Learning on Label Semantics and Enriched Representation for Relation Extraction

Authors: Zhenyu Zhou, Qinghua Zhang, Fan Zhao

Published in: Cognitive Computation

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Sentence-level relation extraction is a technique for extracting factual information about relationships between entities from a sentence. However, the customary method overlooks the semantic information conveyed by the label itself, thereby compromising the efficacy of rare types. Furthermore, there is a growing interest in exploring the use of textual information as a crucial resource to enhance RE models for more effectiveness. To address these two issues, CLERE (Contrastive Learning and Enriched Representation for Relation Extraction) based on contrastive learning and enriched representation of context is proposed. Firstly, by contrastive learning to incorporate semantic information of labels, CLERE is able to effectively convey and exploit the underlying semantics of various sample categories. Thereby enhancing its semantics understanding and classification capabilities, the issue of misclassification due to data imbalance is alleviated. Secondly, both semantics of context and positional information of tagged entities are enhanced by employing weighted layer pooling on pre-trained language models, which improves the representation of context and entity mentions. Experiments are conducted on three public dataset to authenticate the effectiveness of CLERE. The results demonstrate that the proposed model outperforms existing mainstream baseline methods significantly.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Nguyen T, Grishman R. Event detection and domain adaptation with convolutional neural networks. In: Proceedings of the 53rd annual meeting of the association for computational linguistics and the 7th international joint conference on natural language processing. 2015. p. 365-71. Nguyen T, Grishman R. Event detection and domain adaptation with convolutional neural networks. In: Proceedings of the 53rd annual meeting of the association for computational linguistics and the 7th international joint conference on natural language processing. 2015. p. 365-71.
2.
go back to reference Lample G, Ballesteros M, Subramanian S, Kawakami K, Dyer C. Neural architectures for named entity recognition. In: Proceedings of the 2016 conference of the north american chapter of the association for computational linguistics: human language technologies. 2016. p. 260-70. Lample G, Ballesteros M, Subramanian S, Kawakami K, Dyer C. Neural architectures for named entity recognition. In: Proceedings of the 2016 conference of the north american chapter of the association for computational linguistics: human language technologies. 2016. p. 260-70.
3.
go back to reference Zha E, Zeng D, Lin M, Shen Y. CEPTNER: contrastive learning enhanced prototypical network for two-stage few-shot named entity recognition. Knowl-Based Syst. 2024;295:111730.CrossRef Zha E, Zeng D, Lin M, Shen Y. CEPTNER: contrastive learning enhanced prototypical network for two-stage few-shot named entity recognition. Knowl-Based Syst. 2024;295:111730.CrossRef
4.
go back to reference Chen W, Hong D, Zheng C. Learning knowledge graph embedding with entity descriptions based on LSTM networks. In: 2020 IEEE International Symposium on Product Compliance Engineering-Asia (ISPCE-CN). 2020. p. 1-7. Chen W, Hong D, Zheng C. Learning knowledge graph embedding with entity descriptions based on LSTM networks. In: 2020 IEEE International Symposium on Product Compliance Engineering-Asia (ISPCE-CN). 2020. p. 1-7.
5.
go back to reference Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez A, et al. Attention is all you need. In: Advances in neural information processing systems, vol. 30. 2017. p. 6000–10. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez A, et al. Attention is all you need. In: Advances in neural information processing systems, vol. 30. 2017. p. 6000–10.
6.
go back to reference Devlin J, Chang M, Lee K, Toutanova K. BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies. 2019. p. 4171-86. Devlin J, Chang M, Lee K, Toutanova K. BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies. 2019. p. 4171-86.
7.
8.
go back to reference Zhou W, Chen M. An improved baseline for sentence-level relation extraction. In: Proceedings of the 2nd conference of the Asia-Pacific chapter of the association for computational linguistics and the 12th international joint conference on natural language processing. 2022. p. 161-8. Zhou W, Chen M. An improved baseline for sentence-level relation extraction. In: Proceedings of the 2nd conference of the Asia-Pacific chapter of the association for computational linguistics and the 12th international joint conference on natural language processing. 2022. p. 161-8.
9.
go back to reference Wang X, Gao T, Zhu Z, Zhang Z, Liu Z, Li J, et al. KEPLER: a unified model for knowledge embedding and pre-trained language representation. Trans Assoc Comput Linguistics. 2021;9:176–94.CrossRef Wang X, Gao T, Zhu Z, Zhang Z, Liu Z, Li J, et al. KEPLER: a unified model for knowledge embedding and pre-trained language representation. Trans Assoc Comput Linguistics. 2021;9:176–94.CrossRef
10.
go back to reference Baldini Soares L, FitzGerald N, Ling J, Kwiatkowski T. Matching the blanks: distributional similarity for relation learning. In: Proceedings of the 57th annual meeting of the association for computational linguistics. 2019. p. 2895-2905. Baldini Soares L, FitzGerald N, Ling J, Kwiatkowski T. Matching the blanks: distributional similarity for relation learning. In: Proceedings of the 57th annual meeting of the association for computational linguistics. 2019. p. 2895-2905.
11.
go back to reference Wu S, He Y. Enriching pre-trained language model with entity information for relation classification In: Proceedings of the 28th ACM international conference on information and knowledge management. 2019. p. 2361-64. Wu S, He Y. Enriching pre-trained language model with entity information for relation classification In: Proceedings of the 28th ACM international conference on information and knowledge management. 2019. p. 2361-64.
12.
go back to reference Li Z, Sharaf M, Sitbon L, Du X, Zhou X. CoRE: a context-aware relation extraction method for relation completion. In: 2023 Third International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT). 2023. p. 1-4. Li Z, Sharaf M, Sitbon L, Du X, Zhou X. CoRE: a context-aware relation extraction method for relation completion. In: 2023 Third International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT). 2023. p. 1-4.
13.
go back to reference Huang J, Li B, Xu J, Chen M. Unified semantic typing with meaningful label inference. In: Proceedings of the 2022 conference of the north american chapter of the association for computational linguistics: human language technologies. 2022. p. 2642-54. Huang J, Li B, Xu J, Chen M. Unified semantic typing with meaningful label inference. In: Proceedings of the 2022 conference of the north american chapter of the association for computational linguistics: human language technologies. 2022. p. 2642-54.
14.
go back to reference Reimers N, Gurevych I. Sentence-BERT: sentence embeddings using siamese BERT-networks. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). 2019. p. 3980-90. Reimers N, Gurevych I. Sentence-BERT: sentence embeddings using siamese BERT-networks. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). 2019. p. 3980-90.
15.
go back to reference Nayak Y, Majumder N, Goyal P, Poria S. Deep neural approaches to relation triplets extraction: a comprehensive survey. Cogn Comput. 2021;5(13):1215–32.CrossRef Nayak Y, Majumder N, Goyal P, Poria S. Deep neural approaches to relation triplets extraction: a comprehensive survey. Cogn Comput. 2021;5(13):1215–32.CrossRef
16.
go back to reference Mondal A, Cambria E, Das D, Hussain A, Bandyopadhyay S. Relation extraction of medical concepts using categorization and sentiment analysis. Cogn Comput. 2018;10:670–85.CrossRef Mondal A, Cambria E, Das D, Hussain A, Bandyopadhyay S. Relation extraction of medical concepts using categorization and sentiment analysis. Cogn Comput. 2018;10:670–85.CrossRef
17.
go back to reference Peng H, Gao T, Han X, Lin Y, Li P, Liu Z, et al. Learning from context or names? An empirical study on neural relation extraction. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 2020. p. 3661-72. Peng H, Gao T, Han X, Lin Y, Li P, Liu Z, et al. Learning from context or names? An empirical study on neural relation extraction. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 2020. p. 3661-72.
18.
go back to reference Hu M, Zhang C, Ma F, Liu C, Wen L, Yu P. Semi-supervised relation extraction via incremental meta self-training. In: Findings of the association for computational linguistics: EMNLP 2021. 2021. p. 487-96. Hu M, Zhang C, Ma F, Liu C, Wen L, Yu P. Semi-supervised relation extraction via incremental meta self-training. In: Findings of the association for computational linguistics: EMNLP 2021. 2021. p. 487-96.
19.
go back to reference Gao T, Yao H, Chen D. SimCSE: simple contrastive learning of sentence embeddings. In: Proceedings of the 2021 conference on empirical methods in natural language processing. 2021. p. 6894-910. Gao T, Yao H, Chen D. SimCSE: simple contrastive learning of sentence embeddings. In: Proceedings of the 2021 conference on empirical methods in natural language processing. 2021. p. 6894-910.
20.
go back to reference Khosla P, Teterwak P, Wang C, Sarna A, Tian Y, Isola P, et al. Supervised contrastive learning. In: Proceedings of the 34th international conference on neural information processing systems, vol. 33. 2020. p. 18661-73. Khosla P, Teterwak P, Wang C, Sarna A, Tian Y, Isola P, et al. Supervised contrastive learning. In: Proceedings of the 34th international conference on neural information processing systems, vol. 33. 2020. p. 18661-73.
21.
go back to reference Nguyen D, Matsuo Y, Ishizuka M. Subtree mining for relation extraction from Wikipedia. In: Human language technologies 2007: the conference of the North American chapter of the association for computational linguistics. 2007. p. 125-28. Nguyen D, Matsuo Y, Ishizuka M. Subtree mining for relation extraction from Wikipedia. In: Human language technologies 2007: the conference of the North American chapter of the association for computational linguistics. 2007. p. 125-28.
22.
go back to reference Liu C, Sun W, Chao W, Che W. Convolution neural network for relation extraction. Advan Data Mining Appl. 2013;8347:231–42.CrossRef Liu C, Sun W, Chao W, Che W. Convolution neural network for relation extraction. Advan Data Mining Appl. 2013;8347:231–42.CrossRef
23.
go back to reference Zeng D, Liu K, Lai S, Zhou G, Zhao J. Relation classification via convolutional deep neural network. In: Proceedings of COLING 2014, the 25th international conference on computational linguistics. 2014. p. 2335-44. Zeng D, Liu K, Lai S, Zhou G, Zhao J. Relation classification via convolutional deep neural network. In: Proceedings of COLING 2014, the 25th international conference on computational linguistics. 2014. p. 2335-44.
24.
go back to reference Nguyen T, Grishman R. Relation extraction: perspective from convolutional neural networks. In: Proceedings of the 1st workshop on vector space modeling for natural language processing. 2015. p. 39–48. Nguyen T, Grishman R. Relation extraction: perspective from convolutional neural networks. In: Proceedings of the 1st workshop on vector space modeling for natural language processing. 2015. p. 39–48.
25.
go back to reference Zhang R, Meng F, Zhou Y, Liu B. Relation classification via recurrent neural network with attention and tensor layers. Big Data Mining Anal. 2018;3(1):234–44.CrossRef Zhang R, Meng F, Zhou Y, Liu B. Relation classification via recurrent neural network with attention and tensor layers. Big Data Mining Anal. 2018;3(1):234–44.CrossRef
26.
go back to reference Xu Y, Mou L, Li G, Chen Y, Peng H, Jin Z. Classifying relations via long short term memory networks along shortest dependency paths. In: Proceedings of the 2015 conference on empirical methods in natural language processing. 2015. p. 1785-94. Xu Y, Mou L, Li G, Chen Y, Peng H, Jin Z. Classifying relations via long short term memory networks along shortest dependency paths. In: Proceedings of the 2015 conference on empirical methods in natural language processing. 2015. p. 1785-94.
27.
go back to reference Xu S, Sun S, Zhang Z, Xu F, Liu J. BERT gated multi-window attention network for relation extraction. Neurocomputing. 2022;492:516–29.CrossRef Xu S, Sun S, Zhang Z, Xu F, Liu J. BERT gated multi-window attention network for relation extraction. Neurocomputing. 2022;492:516–29.CrossRef
28.
go back to reference Peters M, Neumann M, Logan R, Schwartz R, Joshi V, Singh S, et al. Knowledge enhanced contextual word representations. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). 2019. p. 43-54. Peters M, Neumann M, Logan R, Schwartz R, Joshi V, Singh S, et al. Knowledge enhanced contextual word representations. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). 2019. p. 43-54.
29.
go back to reference Yamada I, Asai A, Shindo H, Takeda H, Matsumoto Y. LUKE: deep contextualized entity representations with entity-aware self-attention. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 2020. p. 6442-54. Yamada I, Asai A, Shindo H, Takeda H, Matsumoto Y. LUKE: deep contextualized entity representations with entity-aware self-attention. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 2020. p. 6442-54.
31.
go back to reference Hadsell R, Chopra S, LeCun Y. Dimensionality reduction by learning an invariant mapping. In: 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06). 2006. p. 1735-42. Hadsell R, Chopra S, LeCun Y. Dimensionality reduction by learning an invariant mapping. In: 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06). 2006. p. 1735-42.
32.
go back to reference Liu J, Liu J, Wang Q, Wang J, Wu W, Xian Y, et al. RankCSE: unsupervised sentence representations learning via learning to rank. In: Proceedings of the 61st annual meeting of the association for computational linguistics. 2023. p. 13785-802. Liu J, Liu J, Wang Q, Wang J, Wu W, Xian Y, et al. RankCSE: unsupervised sentence representations learning via learning to rank. In: Proceedings of the 61st annual meeting of the association for computational linguistics. 2023. p. 13785-802.
33.
go back to reference Gunel B, Du J, Conneau A, Stoyanov V. Supervised contrastive learning for pre-trained language model fine-tuning. 2021. arXiv:2011.01403 Gunel B, Du J, Conneau A, Stoyanov V. Supervised contrastive learning for pre-trained language model fine-tuning. 2021. arXiv:​2011.​01403
34.
go back to reference Chen T, Shi H, Tang S, Chen Z, Wu F, Zhuang Y. CIL: contrastive instance learning framework for distantly supervised relation extraction. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing. 2021. p. 6191–200. Chen T, Shi H, Tang S, Chen Z, Wu F, Zhuang Y. CIL: contrastive instance learning framework for distantly supervised relation extraction. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing. 2021. p. 6191–200.
35.
go back to reference Zhu X, Meng Q, Ding B, Gu L, Yang Y. Weighted pooling for image recognition of deep convolutional neural networks. Clust Comput. 2019;22(Suppl 4):9371–83.CrossRef Zhu X, Meng Q, Ding B, Gu L, Yang Y. Weighted pooling for image recognition of deep convolutional neural networks. Clust Comput. 2019;22(Suppl 4):9371–83.CrossRef
36.
go back to reference Zhang Y, Zhong V, Chen D, Angeli G, Manning C. Position-aware attention and supervised data improve slot filling. In: Proceedings of the 2017 conference on empirical methods in natural language processing. 2017. p. 33-45. Zhang Y, Zhong V, Chen D, Angeli G, Manning C. Position-aware attention and supervised data improve slot filling. In: Proceedings of the 2017 conference on empirical methods in natural language processing. 2017. p. 33-45.
37.
go back to reference Alt C, Gabryszak A, Hennig L. TACRED revisited: a thorough evaluation of the TACRED relation extraction task. In: Proceedings of the 58th annual meeting of the association for computational linguistics. 2020. p. 1558-69. Alt C, Gabryszak A, Hennig L. TACRED revisited: a thorough evaluation of the TACRED relation extraction task. In: Proceedings of the 58th annual meeting of the association for computational linguistics. 2020. p. 1558-69.
38.
go back to reference Stoica G, Platanios E, Poczos B. Re-TACRED: addressing shortcomings of the TACRED dataset. In: Proceedings of the AAAI conference on artificial intelligence, vol. 35. 2021. p. 13843-50. Stoica G, Platanios E, Poczos B. Re-TACRED: addressing shortcomings of the TACRED dataset. In: Proceedings of the AAAI conference on artificial intelligence, vol. 35. 2021. p. 13843-50.
39.
go back to reference Zhang Y, Qi P, Manning C. Graph convolution over pruned dependency trees improves relation extraction. In: Proceedings of the 2018 conference on empirical methods in natural language processing. 2018. p. 2205-15. Zhang Y, Qi P, Manning C. Graph convolution over pruned dependency trees improves relation extraction. In: Proceedings of the 2018 conference on empirical methods in natural language processing. 2018. p. 2205-15.
41.
go back to reference Joshi M, Chen D, Liu Y, Weld D, Zettlemoyer L, et al. SpanBERT: improving pre-training by representing and predicting spans. Trans Assoc Comput Linguistics. 2020;8:64–77.CrossRef Joshi M, Chen D, Liu Y, Weld D, Zettlemoyer L, et al. SpanBERT: improving pre-training by representing and predicting spans. Trans Assoc Comput Linguistics. 2020;8:64–77.CrossRef
42.
go back to reference Yamamoto Y, Matsuzaki T. Absolute position embedding learns sinusoid-like waves for attention based on relative position. In: Proceedings of the 2023 conference on empirical methods in natural language processing. 2023. p. 15-28. Yamamoto Y, Matsuzaki T. Absolute position embedding learns sinusoid-like waves for attention based on relative position. In: Proceedings of the 2023 conference on empirical methods in natural language processing. 2023. p. 15-28.
43.
go back to reference Klein T, Nabi M. miCSE: mutual information contrastive learning for low-shot sentence embeddings. In: Proceedings of the 61st annual meeting of the association for computational linguistics. 2023. p. 6159-77. Klein T, Nabi M. miCSE: mutual information contrastive learning for low-shot sentence embeddings. In: Proceedings of the 61st annual meeting of the association for computational linguistics. 2023. p. 6159-77.
44.
go back to reference Zhuang J, Jing X, Jia X. Mining negative samples on contrastive learning via curricular weighting strategy. Inf Sci. 2024; 668:120534. Zhuang J, Jing X, Jia X. Mining negative samples on contrastive learning via curricular weighting strategy. Inf Sci. 2024; 668:120534.
45.
go back to reference Wang T, Chen L, Zhu X, Lee Y, Gao J. Weighted contrastive learning with false negative control to help long-tailed product classification. In: Proceedings of the 61st annual meeting of the association for computational linguistics. 2023. p. 574-80. Wang T, Chen L, Zhu X, Lee Y, Gao J. Weighted contrastive learning with false negative control to help long-tailed product classification. In: Proceedings of the 61st annual meeting of the association for computational linguistics. 2023. p. 574-80.
Metadata
Title
Scrutinizing Label: Contrastive Learning on Label Semantics and Enriched Representation for Relation Extraction
Authors
Zhenyu Zhou
Qinghua Zhang
Fan Zhao
Publication date
22-08-2024
Publisher
Springer US
Published in
Cognitive Computation
Print ISSN: 1866-9956
Electronic ISSN: 1866-9964
DOI
https://doi.org/10.1007/s12559-024-10338-5

Premium Partner