Skip to main content

2022 | OriginalPaper | Buchkapitel

A Topical Approach to Capturing Customer Insight Dynamics in Social Media

verfasst von : Miguel Palencia-Olivar

Erschienen in: Advances in Information Retrieval

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

With the emergence of the internet, customers have become far more than mere consumers: they are now opinion makers. As such, they share their experience of goods, services, brands, and retailers. People interested in a certain product often reach for these opinions on all kinds of channels with different structures, from forums to microblogging platforms. On these platforms, topics about almost everything proliferate, and can become viral for a certain time before they begin stagnating, or extinguishing. The amount of data is massive, and the data acquisition processes frequently involve web scraping. Even if basic parsing, cleaning, and standardization exist, the variability of noise create the need for ad-hoc tools. All these elements make it difficult to extract customer insights from the internet. To address these issues, I propose to devise time-dynamic, nonparametric neural-based topic models that take topic, document and word linking into account. I also want to extract opinions accordingly with multilingual contexts, all the while making my tools relevant for pretreatment improvement. Last but not least, I want to devise a proper way of evaluating models so as to assess all their aspects.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
My work solely focuses on the insights per-se, not the emitters, and only includes corpus-related information.
 
2
English, french, spanish, italian, german, and dutch.
 
3
As my work is both statistical and computer-science related, I wanted an exhaustive methodology that could unite both fields as much as possible, with as much emphasis on theoretical concerns as on practical concerns.
 
4
Except that, words in a given language are much more likely to appear within contexts in the same language.
 
5
Language detection is out of the scope of this project, so I either rely on datasets’ existing annotations or use off-the-shelf tools.
 
Literatur
1.
Zurück zum Zitat Batmanghelich, K., et al.: Nonparametric spherical topic modeling with word embeddings. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 537–542. Association for Computational Linguistics, Berlin, August 2016. https://doi.org/10.18653/v1/P16-2087 Batmanghelich, K., et al.: Nonparametric spherical topic modeling with word embeddings. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 537–542. Association for Computational Linguistics, Berlin, August 2016. https://​doi.​org/​10.​18653/​v1/​P16-2087
2.
Zurück zum Zitat Blei, D., et al.: Latent Dirichlet allocation. J. Mach. Learn. Res. 3, 993–1022 (2003)MATH Blei, D., et al.: Latent Dirichlet allocation. J. Mach. Learn. Res. 3, 993–1022 (2003)MATH
3.
Zurück zum Zitat Blei, D.M.: Build, compute, critique, repeat: data analysis with latent variable models. Ann. Rev. Stat. Appl. 1(1), 203–232 (2014)CrossRef Blei, D.M.: Build, compute, critique, repeat: data analysis with latent variable models. Ann. Rev. Stat. Appl. 1(1), 203–232 (2014)CrossRef
5.
Zurück zum Zitat Blei, D.M., et al.: Variational inference: a review for statisticians. J. Am. Stat. Assoc. 112(518), 859–877 (2017)MathSciNetCrossRef Blei, D.M., et al.: Variational inference: a review for statisticians. J. Am. Stat. Assoc. 112(518), 859–877 (2017)MathSciNetCrossRef
6.
Zurück zum Zitat Boyd-Graber, J., et al.: Applications of topic models. Found. Trends Inf. Retrieval 11, 143–296 (2017) Boyd-Graber, J., et al.: Applications of topic models. Found. Trends Inf. Retrieval 11, 143–296 (2017)
7.
Zurück zum Zitat Chang, J., et al.: Reading tea leaves: how humans interpret topic models. In: Bengio, Y., Schuurmans, D., Lafferty, J., Williams, C., Culotta, A. (eds.) Advances in Neural Information Processing Systems, vol. 22. Curran Associates, Inc. (2009) Chang, J., et al.: Reading tea leaves: how humans interpret topic models. In: Bengio, Y., Schuurmans, D., Lafferty, J., Williams, C., Culotta, A. (eds.) Advances in Neural Information Processing Systems, vol. 22. Curran Associates, Inc. (2009)
8.
Zurück zum Zitat Das, R., et al.: Gaussian LDA for topic models with word embeddings. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 795–804. Association for Computational Linguistics, Beijing, July 2015 Das, R., et al.: Gaussian LDA for topic models with word embeddings. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 795–804. Association for Computational Linguistics, Beijing, July 2015
9.
Zurück zum Zitat Dieng, A.B., et al.: TopicRNN: a recurrent neural network with long-range semantic dependency. arXiv:1611.01702 [cs, stat], February 2017 Dieng, A.B., et al.: TopicRNN: a recurrent neural network with long-range semantic dependency. arXiv:​1611.​01702 [cs, stat], February 2017
10.
Zurück zum Zitat Dieng, A.B., et al.: Topic modeling in embedding spaces. Trans. Assoc. Comput. Linguistics 8, 439–453 (2020)MathSciNetCrossRef Dieng, A.B., et al.: Topic modeling in embedding spaces. Trans. Assoc. Comput. Linguistics 8, 439–453 (2020)MathSciNetCrossRef
11.
Zurück zum Zitat Dieng, A.B., Ruiz, F.J.R., Blei, D.M.: The dynamic embedded topic model. CoRR abs/1907.05545 (2019) Dieng, A.B., Ruiz, F.J.R., Blei, D.M.: The dynamic embedded topic model. CoRR abs/1907.05545 (2019)
12.
Zurück zum Zitat Ding, R., Nallapati, R., Xiang, B.: Coherence-aware neural topic modeling. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. pp. 830–836. Association for Computational Linguistics, Brussels, Belgium, October–November 2018 Ding, R., Nallapati, R., Xiang, B.: Coherence-aware neural topic modeling. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. pp. 830–836. Association for Computational Linguistics, Brussels, Belgium, October–November 2018
13.
Zurück zum Zitat Figurnov, M., et al.: Implicit reparameterization gradients. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS 2018, pp. 439–450. Curran Associates Inc., Red Hook (2018) Figurnov, M., et al.: Implicit reparameterization gradients. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS 2018, pp. 439–450. Curran Associates Inc., Red Hook (2018)
14.
Zurück zum Zitat Gelman, A., Meng, X.L., Stern, H.: Posterior predictive assessment of model fitness via realized discrepancies, p. 76 Gelman, A., Meng, X.L., Stern, H.: Posterior predictive assessment of model fitness via realized discrepancies, p. 76
16.
Zurück zum Zitat Kingma, D.P., Welling, M.: Auto-encoding variational bayes. CoRR (2014) Kingma, D.P., Welling, M.: Auto-encoding variational bayes. CoRR (2014)
17.
Zurück zum Zitat Miao, Y., et al.: Neural variational inference for text processing. In: Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48, ICML 2016, pp. 1727–1736. JMLR.org (2016) Miao, Y., et al.: Neural variational inference for text processing. In: Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48, ICML 2016, pp. 1727–1736. JMLR.org (2016)
18.
Zurück zum Zitat Mikolov, T., et al.: Efficient estimation of word representations in vector space. In: Proceedings of Workshop at ICLR 2013, January 2013 Mikolov, T., et al.: Efficient estimation of word representations in vector space. In: Proceedings of Workshop at ICLR 2013, January 2013
19.
Zurück zum Zitat Nalisnick, E.T., Smyth, P.: Stick-breaking variational autoencoders. In: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, 24–26 April, 2017, Conference Track Proceedings. OpenReview.net (2017) Nalisnick, E.T., Smyth, P.: Stick-breaking variational autoencoders. In: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, 24–26 April, 2017, Conference Track Proceedings. OpenReview.net (2017)
20.
Zurück zum Zitat Ning, X., et al.: Nonparametric topic modeling with neural inference. Neurocomputing 399, 296–306 (2020)CrossRef Ning, X., et al.: Nonparametric topic modeling with neural inference. Neurocomputing 399, 296–306 (2020)CrossRef
22.
Zurück zum Zitat Rezende, D.J., Mohamed, S., Wierstra, D.: Stochastic backpropagation and approximate inference in deep generative models. In: Xing, E.P., Jebara, T. (eds.) Proceedings of the 31st International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 32, pp. 1278–1286. PMLR, Bejing, 22–24 June 2014 Rezende, D.J., Mohamed, S., Wierstra, D.: Stochastic backpropagation and approximate inference in deep generative models. In: Xing, E.P., Jebara, T. (eds.) Proceedings of the 31st International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 32, pp. 1278–1286. PMLR, Bejing, 22–24 June 2014
23.
Zurück zum Zitat Srivastava, A., Sutton, C.: Autoencoding Variational Inference For Topic Models, p. 12 (2017) Srivastava, A., Sutton, C.: Autoencoding Variational Inference For Topic Models, p. 12 (2017)
25.
Zurück zum Zitat Xun, G., et al.: A correlated topic model using word embeddings. In: Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, IJCAI 2017, pp. 4207–4213 (2017) Xun, G., et al.: A correlated topic model using word embeddings. In: Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, IJCAI 2017, pp. 4207–4213 (2017)
26.
Zurück zum Zitat Yan, X., et al.: A biterm topic model for short texts. In: WWW 2013 - Proceedings of the 22nd International Conference on World Wide Web, pp. 1445–1456 (2013) Yan, X., et al.: A biterm topic model for short texts. In: WWW 2013 - Proceedings of the 22nd International Conference on World Wide Web, pp. 1445–1456 (2013)
Metadaten
Titel
A Topical Approach to Capturing Customer Insight Dynamics in Social Media
verfasst von
Miguel Palencia-Olivar
Copyright-Jahr
2022
DOI
https://doi.org/10.1007/978-3-030-99739-7_64