Skip to main content

2023 | OriginalPaper | Buchkapitel

Chinese Medical Text Classification with RoBERTa

verfasst von : Fengquan Cai, Hui Ye

Erschienen in: Biomedical and Computational Biology

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Many existed Chinese text classification solutions are successful, but the gap is that they are also limited by the models applied by themselves, so it’s available to consider a solution for advancing the Chinese text classification performance, especially in TCM (Traditional Chinese Medicine) text classification task. Assembled by Encoder element and Decoder element, Transformer and others X-former models have shown an outstanding performance in different NLP tasks, and among them BERT has succeeded in text representation and text classification tasks, but it has the possibility to be improved. Here we show our solution and experiment. In many NLP tasks, RoBERTa, which is based on BERT, has s a state-of-the-art performance than BERT. The classified sample data is selected from TCM workbench and tokenized by the Tokenizer we build based on pretrained RoBERTa, which was processed by RoBERTa_TCM, the RoBERTa model fine-tuned with our own data. In order to evaluate the vectorization performance and text classification performance of our Tokenizer-RoBERTa_TCM solution, we select some wild-range-applied language model: Word2Vec, LSTM, Bi-LSTM, contributing 4 baselines: Word2Vec-LSTM, Word2Vec-BiLSTM, Tokenizer-LSTM, Tokenizer-BiLSTM. We find out that the Tokenizer-RoBERTa_TCM model has shown a state-of-the-art classification ability with 90.88% average precision, 91.05% average recall and 90.72% average F1. All of them were the highest results among the baselines. It means that compared to regular text classification models (LSTM, Bi-LSTM, etc.), our RoBERTa_TCM model has an obvious improvement. This solution has the potential application research value in the text classification of TCM text.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Zhou, Y.: Research on medical text analysis and mining technology based on machine learning. Beijing Jiaotong University (2019) Zhou, Y.: Research on medical text analysis and mining technology based on machine learning. Beijing Jiaotong University (2019)
3.
Zurück zum Zitat Gong, W., Wei, X.: News text classification methods based on BiLSTM-CNN model. Comput. Knowl. Technol. 17(21), 105–107 (2021) Gong, W., Wei, X.: News text classification methods based on BiLSTM-CNN model. Comput. Knowl. Technol. 17(21), 105–107 (2021)
4.
Zurück zum Zitat Qu, Q., Kan, H.: Named entity recognition of Chinese medical text based on Bert-BiLSTM-CRF. Electr. Design Eng. 29(19), 40–43, 48 (2021) Qu, Q., Kan, H.: Named entity recognition of Chinese medical text based on Bert-BiLSTM-CRF. Electr. Design Eng. 29(19), 40–43, 48 (2021)
5.
Zurück zum Zitat Yan, L., Zhu, X., Chen, X.: Emotional Classification algorithm of comment text based on two-channel fusion and BiLSTM-attention. J. Univ. Shanghai Sci. Technol. 43(06), 597–605 (2021) Yan, L., Zhu, X., Chen, X.: Emotional Classification algorithm of comment text based on two-channel fusion and BiLSTM-attention. J. Univ. Shanghai Sci. Technol. 43(06), 597–605 (2021)
6.
Zurück zum Zitat Ye, H., Du, L., Lin, S., et al.: Extraction and classification of TCM medical records based on BERT and Bi-LSTM With attention mechanism. In: 2020 IEEE International Conference on Bioinformatics and Biomedicine, BIBM, pp. 1626–1631. IEEE (2020) Ye, H., Du, L., Lin, S., et al.: Extraction and classification of TCM medical records based on BERT and Bi-LSTM With attention mechanism. In: 2020 IEEE International Conference on Bioinformatics and Biomedicine, BIBM, pp. 1626–1631. IEEE (2020)
7.
Zurück zum Zitat Devlin, J., Chang, M., W., et al.: BERT: pre-training of bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018) Devlin, J., Chang, M., W., et al.: BERT: pre-training of bidirectional transformers for language understanding. arXiv preprint arXiv:​1810.​04805 (2018)
8.
Zurück zum Zitat Vaswani, A., Shazeer, N., Parmar, N., et al.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017) Vaswani, A., Shazeer, N., Parmar, N., et al.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)
9.
Zurück zum Zitat Lan, Z., Chen, M., Goodman, S., Gimpel, K, et al.: ALBERT: a lite BERT for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942 (2019) Lan, Z., Chen, M., Goodman, S., Gimpel, K, et al.: ALBERT: a lite BERT for self-supervised learning of language representations. arXiv preprint arXiv:​1909.​11942 (2019)
10.
Zurück zum Zitat Sanh, V., Debut, L., Chaumond, J, et al.: DistillBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 (2019) Sanh, V., Debut, L., Chaumond, J, et al.: DistillBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:​1910.​01108 (2019)
11.
Zurück zum Zitat Lample, G., Conneau, A.: Cross-lingual language model pretraining. Adv. Neural Inf. Process. Syst. 32 (2019) Lample, G., Conneau, A.: Cross-lingual language model pretraining. Adv. Neural Inf. Process. Syst. 32 (2019)
13.
Zurück zum Zitat Clark, K., Luong, M., T., et al.: ELECTRA: pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555 (2020) Clark, K., Luong, M., T., et al.: ELECTRA: pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:​2003.​10555 (2020)
14.
Zurück zum Zitat Radford, A., Narasimhan, K., Salimans, T., et al.: Improving language understanding by generative pre-training (2018) Radford, A., Narasimhan, K., Salimans, T., et al.: Improving language understanding by generative pre-training (2018)
15.
Zurück zum Zitat Radford, A., Wu, J., Child, R., et al.: Language models are unsupervised multitask learners. Open AI Blog 1(8), 9 (2019) Radford, A., Wu, J., Child, R., et al.: Language models are unsupervised multitask learners. Open AI Blog 1(8), 9 (2019)
16.
Zurück zum Zitat Shirish Keskar, N., McCann, B., Varshney, L., et al.: CTRL: a conditional transformer language model for controllable generation. arXiv preprint arXiv:1909.05858 (2019) Shirish Keskar, N., McCann, B., Varshney, L., et al.: CTRL: a conditional transformer language model for controllable generation. arXiv preprint arXiv:​1909.​05858 (2019)
18.
Zurück zum Zitat Dai, Z., Yang, Z., Yang, Y., et al: Transformer-XL: attentive language models beyond a fixed-length context. arXiv preprint arXiv:1901.02860 (2019) Dai, Z., Yang, Z., Yang, Y., et al: Transformer-XL: attentive language models beyond a fixed-length context. arXiv preprint arXiv:​1901.​02860 (2019)
20.
Zurück zum Zitat Liu, Y., Ott, M., Goyal, N., et al.: RoBERTa: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692 (2019) Liu, Y., Ott, M., Goyal, N., et al.: RoBERTa: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:​1907.​11692 (2019)
21.
22.
Zurück zum Zitat Hochreiter, S., Schmidhuber, J., et al.: Long-short term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRefPubMed Hochreiter, S., Schmidhuber, J., et al.: Long-short term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRefPubMed
23.
Zurück zum Zitat Graves, A., Schmidhuber, J.: Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 18(5–6), 602–610 (2005)CrossRefPubMed Graves, A., Schmidhuber, J.: Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 18(5–6), 602–610 (2005)CrossRefPubMed
24.
Zurück zum Zitat Mikolov, T., Chen, K., Corrado, G., et al.: Efficient Estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 (2013) Mikolov, T., Chen, K., Corrado, G., et al.: Efficient Estimation of word representations in vector space. arXiv preprint arXiv:​1301.​3781 (2013)
25.
Metadaten
Titel
Chinese Medical Text Classification with RoBERTa
verfasst von
Fengquan Cai
Hui Ye
Copyright-Jahr
2023
DOI
https://doi.org/10.1007/978-3-031-25191-7_17

Premium Partner