Skip to main content
Erschienen in: Neural Processing Letters 2/2022

01.11.2021

WakaVT: A Sequential Variational Transformer for Waka Generation

verfasst von: Yuka Takeishi, Mingxuan Niu, Jing Luo, Zhong Jin, Xinyu Yang

Erschienen in: Neural Processing Letters | Ausgabe 2/2022

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Poetry generation has long been a challenge for artificial intelligence. In the scope of Japanese poetry generation, many researchers have paid attention to Haiku generation, but few have focused on Waka generation. To further explore the creative potential of natural language generation systems in Japanese poetry creation, we propose a novel Waka generation model, WakaVT, which automatically produces Waka poems given user-specified keywords. Firstly, an additive mask-based approach is presented to satisfy the form constraint. Secondly, the structures of Transformer and variational autoencoder are integrated to enhance the quality of generated content. Specifically, to obtain novelty and diversity, WakaVT employs a sequence of latent variables, which effectively captures word-level variability in Waka data. To improve linguistic quality in terms of fluency, coherence, and meaningfulness, we further propose the fused multilevel self-attention mechanism, which properly models the hierarchical linguistic structure of Waka. To the best of our knowledge, we are the first to investigate Waka generation with models based on Transformer and/or variational autoencoder. Both objective and subjective evaluation results demonstrate that our model outperforms baselines significantly.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
To avoid ambiguity, our study of Waka is limited to Tanka. For an introduction to various types of Waka, see https://​en.​wikipedia.​org/​wiki/​Waka_​(poetry).
 
2
In the Japanese writing system, the dakuten (voicing mark) is used to indicate kana characters supposed to be pronounced voiced. However, it’s not used by ancient poets in the writing of Waka, making it difficult to recognize the pronunciations. For more information, see https://​en.​m.​wikipedia.​org/​wiki/​Dakuten_​and_​handakuten.
 
3
Our code, samples generated by our model and the corresponding comments given by experts are available at https://​github.​com/​nmx-cs/​wakavt.
 
4
Selected from the collection of Bunpo Hyakushu, edited by Emperor Go-Uta.
 
5
Additive masks can also be used to avoid incomplete moraes. However, we have found that most generation could follow the correct pattern even if we don’t consider this problem.
 
9
Among the generated poems used for objective evaluation, the proportions of morae constrained poems generated by TLM, TVAE, WakaVT, and RNN-VAD were 92.7%, 98.2%, 99.5%, and 100%, respectively.
 
10
Unilateral T-test is used for significance test.
 
11
The calculated PPL values of WakaVT and WakaVT + FMSA with this poem as input are 5.39 and 3.15, respectively.
 
12
In this poem, the wind in the sky is a metaphor for lovers meeting.
 
13
In the training set, at least 5 poems created by humans start with .
 
Literatur
2.
Zurück zum Zitat Bojanowski P, Grave E, Joulin A, Mikolov T (2017) Enriching word vectors with subword information. Trans Assoc Comput Linguist 5:135–146CrossRef Bojanowski P, Grave E, Joulin A, Mikolov T (2017) Enriching word vectors with subword information. Trans Assoc Comput Linguist 5:135–146CrossRef
3.
Zurück zum Zitat Chen H, Yi X, Sun M, Li W, Yang C, Guo Z (2019) Sentiment-controllable Chinese poetry generation. In: Proceedings of the twenty-eighth international joint conference on artificial intelligence (IJCAI), pp 4925–4931 Chen H, Yi X, Sun M, Li W, Yang C, Guo Z (2019) Sentiment-controllable Chinese poetry generation. In: Proceedings of the twenty-eighth international joint conference on artificial intelligence (IJCAI), pp 4925–4931
4.
Zurück zum Zitat Colton S, Goodwin J, Veale T (2012) Full-face poetry generation. In: Proceedings of the third international conference on computational creativity (ICCC), pp 95–102 Colton S, Goodwin J, Veale T (2012) Full-face poetry generation. In: Proceedings of the third international conference on computational creativity (ICCC), pp 95–102
5.
Zurück zum Zitat Deng L, Wang J, Liang H, Chen H, Xie Z, Zhuang B, Wang S, Xiao J (2020) An iterative polishing framework based on quality aware masked language model for Chinese poetry generation. In: Proceedings of the AAAI conference on artificial intelligence, pp 7643–7650 Deng L, Wang J, Liang H, Chen H, Xie Z, Zhuang B, Wang S, Xiao J (2020) An iterative polishing framework based on quality aware masked language model for Chinese poetry generation. In: Proceedings of the AAAI conference on artificial intelligence, pp 7643–7650
6.
Zurück zum Zitat Du J, Li W, He Y, Xu R, Bing L, Wang X (2018) Variational autoregressive decoder for neural response generation. In: Proceedings of the 2018 conference on empirical methods in natural language processing (EMNLP), pp 3154–3163 Du J, Li W, He Y, Xu R, Bing L, Wang X (2018) Variational autoregressive decoder for neural response generation. In: Proceedings of the 2018 conference on empirical methods in natural language processing (EMNLP), pp 3154–3163
7.
Zurück zum Zitat Ghazvininejad M, Shi X, Choi Y, Knight K (2016) Generating topical poetry. In: Proceedings of the 2016 conference on empirical methods in natural language processing (EMNLP), pp 1183–1191 Ghazvininejad M, Shi X, Choi Y, Knight K (2016) Generating topical poetry. In: Proceedings of the 2016 conference on empirical methods in natural language processing (EMNLP), pp 1183–1191
8.
Zurück zum Zitat Guo Z, Yi X, Sun M, Li W, Yang C, Liang J, Chen H, Zhang Y, Li R (2019) Jiuge: a human–machine collaborative Chinese classical poetry generation system. In: Proceedings of the 57th annual meeting of the association for computational linguistics: system demonstrations, pp 25–30 Guo Z, Yi X, Sun M, Li W, Yang C, Liang J, Chen H, Zhang Y, Li R (2019) Jiuge: a human–machine collaborative Chinese classical poetry generation system. In: Proceedings of the 57th annual meeting of the association for computational linguistics: system demonstrations, pp 25–30
9.
Zurück zum Zitat He J, Zhou M, Jiang L (2012) Generating Chinese classical poems with statistical machine translation models. In: Proceedings of the twenty-sixth AAAI conference on artificial intelligence, pp 1650–1656 He J, Zhou M, Jiang L (2012) Generating Chinese classical poems with statistical machine translation models. In: Proceedings of the twenty-sixth AAAI conference on artificial intelligence, pp 1650–1656
10.
Zurück zum Zitat Hirota A, Oka N, Araki M, Tanaka K (2018) Haiku generation by Seqgan that divided learning data set. In: Proceedings of the twenty-fourth annual meeting of the association for natural language processing, pp 1292–1295 Hirota A, Oka N, Araki M, Tanaka K (2018) Haiku generation by Seqgan that divided learning data set. In: Proceedings of the twenty-fourth annual meeting of the association for natural language processing, pp 1292–1295
11.
Zurück zum Zitat Hrešková M, Machová K (2017) Haiku poetry generation using interactive evolution vs. poem models. Acta Electrotech Inf 17(1):10–16 Hrešková M, Machová K (2017) Haiku poetry generation using interactive evolution vs. poem models. Acta Electrotech Inf 17(1):10–16
12.
Zurück zum Zitat Ito T, Ono J, Ogata T (2018) Haiku generation using gap techniques. In: Proceedings of the 2018 international conference on artificial intelligence and virtual reality, pp 93–96 Ito T, Ono J, Ogata T (2018) Haiku generation using gap techniques. In: Proceedings of the 2018 international conference on artificial intelligence and virtual reality, pp 93–96
13.
Zurück zum Zitat Ji S, Luo J, Yang X (2020) A comprehensive survey on deep music generation: multi-level representations, algorithms, evaluations, and future directions. arXiv:2011.06801 Ji S, Luo J, Yang X (2020) A comprehensive survey on deep music generation: multi-level representations, algorithms, evaluations, and future directions. arXiv:​2011.​06801
14.
Zurück zum Zitat Keskar NS, McCann B, Varshney LR, Xiong C, Socher R (2019) Ctrl: a conditional transformer language model for controllable generation. arXiv:1909.05858 Keskar NS, McCann B, Varshney LR, Xiong C, Socher R (2019) Ctrl: a conditional transformer language model for controllable generation. arXiv:​1909.​05858
15.
Zurück zum Zitat Kingma DP, Ba J (2015) Adam: a method for stochastic optimization. In: International conference on learning representations, pp 1–15 Kingma DP, Ba J (2015) Adam: a method for stochastic optimization. In: International conference on learning representations, pp 1–15
16.
Zurück zum Zitat Kingma DP, Welling M (2014) Auto-encoding variational bayes. In: International conference on learning representations, pp 1–14 Kingma DP, Welling M (2014) Auto-encoding variational bayes. In: International conference on learning representations, pp 1–14
17.
Zurück zum Zitat Konishi B, Hirota A, Matsuo S, Iehara R, Obara S, Kaga Y, Tsuruda J, Wakigami Y, Kanajiri R, Fukada C, Tanaka K, Oka N (2017) Generation of haiku preferable for ordinary people by Seqgan. In: Information processing society of Japan (IPSJ) Kansai-Branch convention, pp 1–3 Konishi B, Hirota A, Matsuo S, Iehara R, Obara S, Kaga Y, Tsuruda J, Wakigami Y, Kanajiri R, Fukada C, Tanaka K, Oka N (2017) Generation of haiku preferable for ordinary people by Seqgan. In: Information processing society of Japan (IPSJ) Kansai-Branch convention, pp 1–3
18.
Zurück zum Zitat Li J, Song Y, Zhang H, Chen D, Shi S, Zhao D, Yan R (2018) Generating classical Chinese poems via conditional variational autoencoder and adversarial training. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 3890–3900 Li J, Song Y, Zhang H, Chen D, Shi S, Zhao D, Yan R (2018) Generating classical Chinese poems via conditional variational autoencoder and adversarial training. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 3890–3900
19.
Zurück zum Zitat Li P, Zhang H, Liu X, Shi S (2020) Rigid formats controlled text generation. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 742–751 Li P, Zhang H, Liu X, Shi S (2020) Rigid formats controlled text generation. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 742–751
20.
21.
Zurück zum Zitat Liu D, Guo Q, Li W, Lv J (2018) A multi-modal Chinese poetry generation model. In: International joint conference on neural networks (IJCNN), pp 1–8 Liu D, Guo Q, Li W, Lv J (2018) A multi-modal Chinese poetry generation model. In: International joint conference on neural networks (IJCNN), pp 1–8
22.
Zurück zum Zitat Manurung R, Ritchie G, Thompson H (2012) Using genetic algorithms to create meaningful poetic text. J Exp Theor Artif Intell 24(1):43–64CrossRef Manurung R, Ritchie G, Thompson H (2012) Using genetic algorithms to create meaningful poetic text. J Exp Theor Artif Intell 24(1):43–64CrossRef
23.
Zurück zum Zitat Masada T, Takasu A (2018) LDA-based scoring of sequences generated by RNN for automatic Tanka composition. In: International conference on computational science, pp 395–402 Masada T, Takasu A (2018) LDA-based scoring of sequences generated by RNN for automatic Tanka composition. In: International conference on computational science, pp 395–402
24.
Zurück zum Zitat Misztal J, Indurkhya B (2014) Poetry generation system with an emotional personality. In: Proceedings of the fifth international conference on computational creativity (ICCC), pp 72–81 Misztal J, Indurkhya B (2014) Poetry generation system with an emotional personality. In: Proceedings of the fifth international conference on computational creativity (ICCC), pp 72–81
25.
Zurück zum Zitat Netzer Y, Gabay D, Goldberg Y, Elhadad M (2009) Gaiku: generating haiku with word associations norms. In: Proceedings of the workshop on computational approaches to linguistic creativity, pp 32–39 Netzer Y, Gabay D, Goldberg Y, Elhadad M (2009) Gaiku: generating haiku with word associations norms. In: Proceedings of the workshop on computational approaches to linguistic creativity, pp 32–39
26.
Zurück zum Zitat Oliveira HG (2012) Poetryme: a versatile platform for poetry generation. Comput Creat Concept Invent Gen Intell 1:1–21 Oliveira HG (2012) Poetryme: a versatile platform for poetry generation. Comput Creat Concept Invent Gen Intell 1:1–21
27.
Zurück zum Zitat Oliveira HG (2017) A survey on intelligent poetry generation: languages, features, techniques, reutilisation and evaluation. In: Proceedings of the 10th international conference on natural language generation, pp 11–20 Oliveira HG (2017) A survey on intelligent poetry generation: languages, features, techniques, reutilisation and evaluation. In: Proceedings of the 10th international conference on natural language generation, pp 11–20
28.
Zurück zum Zitat Ovalle JEA, Solorio T, Montes-y-Gómez M, González FA (2017) Gated multimodal units for information fusion. In: International conference on learning representations, pp 1–17 Ovalle JEA, Solorio T, Montes-y-Gómez M, González FA (2017) Gated multimodal units for information fusion. In: International conference on learning representations, pp 1–17
29.
Zurück zum Zitat Rzepka R, Araki K (2015) Haiku generator that reads blogs and illustrates them with sounds and images. In: Proceedings of the 24th international joint conference on artificial intelligence (IJCAI), pp 2496–2502 Rzepka R, Araki K (2015) Haiku generator that reads blogs and illustrates them with sounds and images. In: Proceedings of the 24th international joint conference on artificial intelligence (IJCAI), pp 2496–2502
30.
Zurück zum Zitat Schulz P, Aziz W, Cohn T (2018) A stochastic decoder for neural machine translation. In: Proceedings of the 56th annual meeting of the association for computational linguistics (vol 1: Long Papers), pp 1243–1252 Schulz P, Aziz W, Cohn T (2018) A stochastic decoder for neural machine translation. In: Proceedings of the 56th annual meeting of the association for computational linguistics (vol 1: Long Papers), pp 1243–1252
31.
Zurück zum Zitat Shen L, Guo X, Chen M (2020) Compose like humans: jointly improving the coherence and novelty for modern Chinese poetry generation. In: 2020 international joint conference on neural networks (IJCNN), pp 1–8 Shen L, Guo X, Chen M (2020) Compose like humans: jointly improving the coherence and novelty for modern Chinese poetry generation. In: 2020 international joint conference on neural networks (IJCNN), pp 1–8
32.
Zurück zum Zitat Sutton C, McCallum A (2012) An introduction to conditional random fields. Found Trends Mach Learn 4(4):267–373CrossRef Sutton C, McCallum A (2012) An introduction to conditional random fields. Found Trends Mach Learn 4(4):267–373CrossRef
33.
Zurück zum Zitat Tosa N, Obara H, Minoh M (2008) Hitch haiku: an interactive supporting system for composing haiku poem. In: Proceedings of the 7th international conference on entertainment computing, pp 209–216 Tosa N, Obara H, Minoh M (2008) Hitch haiku: an interactive supporting system for composing haiku poem. In: Proceedings of the 7th international conference on entertainment computing, pp 209–216
34.
Zurück zum Zitat Van de Cruys T (2020) Automatic poetry generation from prosaic text. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 2471–2480 Van de Cruys T (2020) Automatic poetry generation from prosaic text. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 2471–2480
35.
Zurück zum Zitat Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. In: Proceedings of the 31st international conference on neural information processing systems, pp 6000–6010 Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. In: Proceedings of the 31st international conference on neural information processing systems, pp 6000–6010
36.
Zurück zum Zitat Wang DZ, He W, Wu H, Wu H, Li W, Wang H, Chen E (2016) Chinese poetry generation with planning based neural network. In: Proceedings of COLING 2016, the 26th international conference on computational linguistics: technical papers, pp 1051–1060 Wang DZ, He W, Wu H, Wu H, Li W, Wang H, Chen E (2016) Chinese poetry generation with planning based neural network. In: Proceedings of COLING 2016, the 26th international conference on computational linguistics: technical papers, pp 1051–1060
37.
Zurück zum Zitat Wong TM, Lee OKS, Chun HWA (2008) Automatic haiku generation using VSM. In: Proceedings of the 7th WSEAS international conference on applied computer and applied computation science (ACACOS’08), pp 318–323 Wong TM, Lee OKS, Chun HWA (2008) Automatic haiku generation using VSM. In: Proceedings of the 7th WSEAS international conference on applied computer and applied computation science (ACACOS’08), pp 318–323
38.
Zurück zum Zitat Wu X, Klyen M, Ito K, Chen Z (2017) Haiku generation using deep neural networks. In: Annual conference of the language processing society, pp 1133–1136 Wu X, Klyen M, Ito K, Chen Z (2017) Haiku generation using deep neural networks. In: Annual conference of the language processing society, pp 1133–1136
39.
Zurück zum Zitat Xu L, Jiang L, Qin C, Wang Z, Du D (2018) How images inspire poems: generating classical Chinese poetry from images with memory networks. In: Proceedings of the AAAI conference on artificial intelligence, pp 5618–5625 Xu L, Jiang L, Qin C, Wang Z, Du D (2018) How images inspire poems: generating classical Chinese poetry from images with memory networks. In: Proceedings of the AAAI conference on artificial intelligence, pp 5618–5625
40.
Zurück zum Zitat Yan R, Jiang H, Lapata M, Lin SD, Lv X, Li X (2013) I, poet: automatic Chinese poetry composition through a generative summarization framework under constrained optimization. In: Proceedings of the twenty-third international joint conference on artificial intelligence, pp 2197–2203 Yan R, Jiang H, Lapata M, Lin SD, Lv X, Li X (2013) I, poet: automatic Chinese poetry composition through a generative summarization framework under constrained optimization. In: Proceedings of the twenty-third international joint conference on artificial intelligence, pp 2197–2203
41.
Zurück zum Zitat Yang C, Sun M, Yi X, Li W (2018) Stylistic Chinese poetry generation via unsupervised style disentanglement. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 3960–3969 Yang C, Sun M, Yi X, Li W (2018) Stylistic Chinese poetry generation via unsupervised style disentanglement. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 3960–3969
42.
Zurück zum Zitat Yang M, Hagiwara M (2016) A text-based automatic Waka generation system using Kansei. Int J Affect Eng 15(2):125–134CrossRef Yang M, Hagiwara M (2016) A text-based automatic Waka generation system using Kansei. Int J Affect Eng 15(2):125–134CrossRef
43.
Zurück zum Zitat Yang X, Lin X, Suo S, Li M (2018) Generating thematic Chinese poetry using conditional variational autoencoders with hybrid decoders. In: Proceedings of the 27th international joint conference on artificial intelligence (IJCAI), pp 4539–4545 Yang X, Lin X, Suo S, Li M (2018) Generating thematic Chinese poetry using conditional variational autoencoders with hybrid decoders. In: Proceedings of the 27th international joint conference on artificial intelligence (IJCAI), pp 4539–4545
44.
Zurück zum Zitat Yi X, Li R, Yang C, Li W, Sun M (2020) Mixpoet: diverse poetry generation via learning controllable mixed latent space. In: Proceedings of the AAAI conference on artificial intelligence, pp 9450–9457 Yi X, Li R, Yang C, Li W, Sun M (2020) Mixpoet: diverse poetry generation via learning controllable mixed latent space. In: Proceedings of the AAAI conference on artificial intelligence, pp 9450–9457
45.
Zurück zum Zitat Zhang J, Feng Y, Wang D, Wang Y, Abel A, Zhang S, Zhang A (2017) Flexible and creative Chinese poetry generation using neural memory. In: Proceedings of the 55th annual meeting of the association for computational linguistics (volume 1: long papers), pp 1364–1373 Zhang J, Feng Y, Wang D, Wang Y, Abel A, Zhang S, Zhang A (2017) Flexible and creative Chinese poetry generation using neural memory. In: Proceedings of the 55th annual meeting of the association for computational linguistics (volume 1: long papers), pp 1364–1373
46.
Zurück zum Zitat Zhang X, Lapata M (2014) Chinese poetry generation with recurrent neural networks. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 670–680 Zhang X, Lapata M (2014) Chinese poetry generation with recurrent neural networks. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 670–680
Metadaten
Titel
WakaVT: A Sequential Variational Transformer for Waka Generation
verfasst von
Yuka Takeishi
Mingxuan Niu
Jing Luo
Zhong Jin
Xinyu Yang
Publikationsdatum
01.11.2021
Verlag
Springer US
Erschienen in
Neural Processing Letters / Ausgabe 2/2022
Print ISSN: 1370-4621
Elektronische ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-021-10654-z

Weitere Artikel der Ausgabe 2/2022

Neural Processing Letters 2/2022 Zur Ausgabe

Neuer Inhalt