Skip to main content
Erschienen in: Discover Computing 4/2020

08.06.2020

Assessing ranking metrics in top-N recommendation

verfasst von: Daniel Valcarce, Alejandro Bellogín, Javier Parapar, Pablo Castells

Erschienen in: Discover Computing | Ausgabe 4/2020

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

The evaluation of recommender systems is an area with unsolved questions at several levels. Choosing the appropriate evaluation metric is one of such important issues. Ranking accuracy is generally identified as a prerequisite for recommendation to be useful. Ranking metrics have been adapted for this purpose from the Information Retrieval field into the recommendation task. In this article, we undertake a principled analysis of the robustness and the discriminative power of different ranking metrics for the offline evaluation of recommender systems, drawing from previous studies in the information retrieval field. We measure the robustness to different sources of incompleteness that arise from the sparsity and popularity biases in recommendation. Among other results, we find that precision provides high robustness while normalized discounted cumulative gain offers the best discriminative power. In dealing with cold users, we also find that the geometric mean is more robust than the arithmetic mean as aggregation function over users.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Anderson, C. (2008). The Long Tail: Why the Future of Business Is Selling Less of More. New York: Hachette Books. Anderson, C. (2008). The Long Tail: Why the Future of Business Is Selling Less of More. New York: Hachette Books.
Zurück zum Zitat Beel, J., & Langer, S. (2015). A comparison of offline evaluations, online evaluations, and user studies in the context of research-paper recommender systems. In S. Kapidakis, C. Mazurek, M. Werla (eds.) Proceedings of the 19th international conference on theory and practice of digital libraries, TPDL ’15 (pp. 153–168). Springer, Cham. https://doi.org/10.1007/978-3-319-24592-8_12. Beel, J., & Langer, S. (2015). A comparison of offline evaluations, online evaluations, and user studies in the context of research-paper recommender systems. In S. Kapidakis, C. Mazurek, M. Werla (eds.) Proceedings of the 19th international conference on theory and practice of digital libraries, TPDL ’15 (pp. 153–168). Springer, Cham. https://​doi.​org/​10.​1007/​978-3-319-24592-8_​12.
Zurück zum Zitat Bennett, J., & Lanning, S (2007). The netflix prize. In Proceedings of the KDD cup workshop 2007 (pp. 3–6). ACM, New York, NY, USA. Bennett, J., & Lanning, S (2007). The netflix prize. In Proceedings of the KDD cup workshop 2007 (pp. 3–6). ACM, New York, NY, USA.
Zurück zum Zitat Blei, D. M., Ng, A. Y., & Jordan, M. I. (2003). Latent Dirichlet allocation. Journal of Machine Learning Research, 3, 993–1022.MATH Blei, D. M., Ng, A. Y., & Jordan, M. I. (2003). Latent Dirichlet allocation. Journal of Machine Learning Research, 3, 993–1022.MATH
Zurück zum Zitat Buckley, C., & Voorhees, E.M (2000). Evaluating evaluation measure stability. In Proceedings of the 23rd annual international ACM SIGIR conference on research and development in information retrieval, SIGIR ’00 (pp. 33–40). ACM, New York, NY, USA. https://doi.org/10.1145/345508.345543. Buckley, C., & Voorhees, E.M (2000). Evaluating evaluation measure stability. In Proceedings of the 23rd annual international ACM SIGIR conference on research and development in information retrieval, SIGIR ’00 (pp. 33–40). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​345508.​345543.
Zurück zum Zitat Buckley, C., & Voorhees, E.M. (2004). Retrieval evaluation with incomplete information. In Proceedings of the 27th annual international ACM SIGIR conference on research and development in information retrieval, SIGIR ’04 (pp. 25–32). ACM, New York, NY, USA. https://doi.org/10.1145/1008992.1009000. Buckley, C., & Voorhees, E.M. (2004). Retrieval evaluation with incomplete information. In Proceedings of the 27th annual international ACM SIGIR conference on research and development in information retrieval, SIGIR ’04 (pp. 25–32). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​1008992.​1009000.
Zurück zum Zitat Büttcher, S., Clarke, C.L.A., Yeung, P.C.K., & Soboroff, I. (2007). Reliable information retrieval evaluation with incomplete and biased judgements. In Proceedings of the 30th annual international ACM SIGIR conference on research and development in information retrieval, SIGIR ’07, (pp. 63–70). ACM, New York, NY, USA. https://doi.org/10.1145/1277741.1277755. Büttcher, S., Clarke, C.L.A., Yeung, P.C.K., & Soboroff, I. (2007). Reliable information retrieval evaluation with incomplete and biased judgements. In Proceedings of the 30th annual international ACM SIGIR conference on research and development in information retrieval, SIGIR ’07, (pp. 63–70). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​1277741.​1277755.
Zurück zum Zitat Cañamares, R., & Castells, P. (2018). Should i follow the crowd? A probabilistic analysis of the effectiveness of popularity in recommender systems. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, SIGIR ’18 (pp. 415–424). ACM, New York, NY, USA. https://doi.org/10.1145/3209978.3210014. Cañamares, R., & Castells, P. (2018). Should i follow the crowd? A probabilistic analysis of the effectiveness of popularity in recommender systems. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, SIGIR ’18 (pp. 415–424). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​3209978.​3210014.
Zurück zum Zitat Chapelle, O., Metlzer, D., Zhang, Y., & Grinspan, P. (2009). Expected reciprocal rank for graded relevance. In Proceedings of the 18th ACM conference on information and knowledge management, CIKM ’09 (p. 621). ACM, New York, NY, USA. https://doi.org/10.1145/1645953.1646033. Chapelle, O., Metlzer, D., Zhang, Y., & Grinspan, P. (2009). Expected reciprocal rank for graded relevance. In Proceedings of the 18th ACM conference on information and knowledge management, CIKM ’09 (p. 621). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​1645953.​1646033.
Zurück zum Zitat Cremonesi, P., Garzotto, F., Negro, S., Papadopoulos, A.V., & Turrin, R. (2011). Looking for “Good” recommendations: A comparative evaluation of recommender systems. In Proceedings of the 13th IFIP TC 13 international conference on human-computer interaction–Part III, INTERACT ’11 (pp. 152–168). Springer, Berlin. https://doi.org/10.1007/978-3-642-23765-2_11. Cremonesi, P., Garzotto, F., Negro, S., Papadopoulos, A.V., & Turrin, R. (2011). Looking for “Good” recommendations: A comparative evaluation of recommender systems. In Proceedings of the 13th IFIP TC 13 international conference on human-computer interaction–Part III, INTERACT ’11 (pp. 152–168). Springer, Berlin. https://​doi.​org/​10.​1007/​978-3-642-23765-2_​11.
Zurück zum Zitat Cremonesi, P., Garzotto, F., & Turrin, R. (2013). User-centric vs. system-centric evaluation of recommender systems. In Proceedings of the 14th IFIP TC 13 international conference on human-computer interaction–Part III, INTERACT ’13 (pp. 334–351). Springer, Berlin. https://doi.org/10.1007/978-3-642-40477-1_21. Cremonesi, P., Garzotto, F., & Turrin, R. (2013). User-centric vs. system-centric evaluation of recommender systems. In Proceedings of the 14th IFIP TC 13 international conference on human-computer interaction–Part III, INTERACT ’13 (pp. 334–351). Springer, Berlin. https://​doi.​org/​10.​1007/​978-3-642-40477-1_​21.
Zurück zum Zitat Cremonesi, P., Koren, Y., & Turrin, R. (2010). Performance of recommender algorithms on top-N recommendation tasks. In Proceedings of the 4th ACM conference on recommender systems, RecSys ’10 (pp. 39–46). ACM, New York, NY, USA. https://doi.org/10.1145/1864708.1864721. Cremonesi, P., Koren, Y., & Turrin, R. (2010). Performance of recommender algorithms on top-N recommendation tasks. In Proceedings of the 4th ACM conference on recommender systems, RecSys ’10 (pp. 39–46). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​1864708.​1864721.
Zurück zum Zitat Efron, B., & Tibshirani, R. J. (1993). An introduction to the bootstrap, monographs on statistics and applied probability. Boca Raton, Florida, USA: Chapman & Hall/CRC.CrossRef Efron, B., & Tibshirani, R. J. (1993). An introduction to the bootstrap, monographs on statistics and applied probability. Boca Raton, Florida, USA: Chapman & Hall/CRC.CrossRef
Zurück zum Zitat Garcin, F., Faltings, B., Donatsch, O., Alazzawi, A., Bruttin, C., & Huber, A. (2014). Offline and online evaluation of news recommender systems at swissinfo.ch. In Proceedings of the 8th ACM conference on recommender systems, RecSys ’14 (pp. 169–176). ACM, New York, NY, USA. https://doi.org/10.1145/2645710.2645745. Garcin, F., Faltings, B., Donatsch, O., Alazzawi, A., Bruttin, C., & Huber, A. (2014). Offline and online evaluation of news recommender systems at swissinfo.ch. In Proceedings of the 8th ACM conference on recommender systems, RecSys ’14 (pp. 169–176). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​2645710.​2645745.
Zurück zum Zitat Gilotte, A., Calauzènes, C., Nedelec, T., Abraham, A., & Dollé, S. (2018). Offline A/B testing for recommender systems. In Proceedings of the 11th ACM international conference on web search and data mining, WSDM ’18 (pp. 198–206). ACM. Gilotte, A., Calauzènes, C., Nedelec, T., Abraham, A., & Dollé, S. (2018). Offline A/B testing for recommender systems. In Proceedings of the 11th ACM international conference on web search and data mining, WSDM ’18 (pp. 198–206). ACM.
Zurück zum Zitat Gini, C. (1912). Variabilità e Mutuabilità. Cuppini: Contributo allo Studio delle Distribuzioni e delle Relazioni Statistiche. Gini, C. (1912). Variabilità e Mutuabilità. Cuppini: Contributo allo Studio delle Distribuzioni e delle Relazioni Statistiche.
Zurück zum Zitat Gruson, A., Chandar, P., Charbuillet, C., McInerney, J., Hansen, S., Tardieu, D., & Carterette, B. (2019). Offline evaluation to make decisions about playlist recommendation. In Proceedings of the 12th ACM international conference on web search and data mining, WSDM ’19 (pp. 420–428). ACM. Gruson, A., Chandar, P., Charbuillet, C., McInerney, J., Hansen, S., Tardieu, D., & Carterette, B. (2019). Offline evaluation to make decisions about playlist recommendation. In Proceedings of the 12th ACM international conference on web search and data mining, WSDM ’19 (pp. 420–428). ACM.
Zurück zum Zitat Hu, Y., Koren, Y., & Volinsky, C. (2008). Collaborative filtering for implicit feedback datasets. In Proceedings of the 2008 eighth IEEE international conference on data mining, ICDM ’08 (pp. 263–272). IEEE, Washington, DC, USA. https://doi.org/10.1109/ICDM.2008.22. Hu, Y., Koren, Y., & Volinsky, C. (2008). Collaborative filtering for implicit feedback datasets. In Proceedings of the 2008 eighth IEEE international conference on data mining, ICDM ’08 (pp. 263–272). IEEE, Washington, DC, USA. https://​doi.​org/​10.​1109/​ICDM.​2008.​22.
Zurück zum Zitat Konstan, J.A., & Adomavicius, G. (2013). Toward identification and adoption of best practices in algorithmic recommender systems research. In Proceedings of the international workshop on reproducibility and replication in recommender systems evaluation, RepSys ’13 (pp. 23–28). ACM, New York, NY, USA. https://doi.org/10.1145/2532508.2532513. Konstan, J.A., & Adomavicius, G. (2013). Toward identification and adoption of best practices in algorithmic recommender systems research. In Proceedings of the international workshop on reproducibility and replication in recommender systems evaluation, RepSys ’13 (pp. 23–28). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​2532508.​2532513.
Zurück zum Zitat Marlin, B. M., Zemel, R. S., Roweis, S., & Slaney, M. (2007). Collaborative filtering and the missing at random assumption. In Proceedings of the 23rd conference on uncertainty in artificial intelligence, UAI’07 (pp. 267–275). Arlington, VA: AUAI Press. Marlin, B. M., Zemel, R. S., Roweis, S., & Slaney, M. (2007). Collaborative filtering and the missing at random assumption. In Proceedings of the 23rd conference on uncertainty in artificial intelligence, UAI’07 (pp. 267–275). Arlington, VA: AUAI Press.
Zurück zum Zitat Matos-Junior, O., Ziviani, N., Botelho, F., Cristo, M., Lacerda, A., & da Silva, A. S. (2012). Using taxonomies for product recommendation. Journal of Information and Data Management, 3(2), 85–100. Matos-Junior, O., Ziviani, N., Botelho, F., Cristo, M., Lacerda, A., & da Silva, A. S. (2012). Using taxonomies for product recommendation. Journal of Information and Data Management, 3(2), 85–100.
Zurück zum Zitat McNee, S. M., Riedl, J. T., & Konstan, J. A. (2006). Being accurate is not enough: How accuracy metrics have hurt recommender systems. In CHI ’06 extended abstracts on human factors in computing systems, CHI EA ’06 (p. 1097). ACM, New York, NY, USA. https://doi.org/10.1145/1125451.1125659. McNee, S. M., Riedl, J. T., & Konstan, J. A. (2006). Being accurate is not enough: How accuracy metrics have hurt recommender systems. In CHI ’06 extended abstracts on human factors in computing systems, CHI EA ’06 (p. 1097). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​1125451.​1125659.
Zurück zum Zitat Parapar, J., Losada, D.E., PresedoQuindimil, M.A., & Barreiro, Á. (2019). Using score distributions to compare statistical significance tests for information retrieval evaluation. Journal of the Association for Information Science and Technology. https://doi.org/10.1002/asi.24203 Parapar, J., Losada, D.E., PresedoQuindimil, M.A., & Barreiro, Á. (2019). Using score distributions to compare statistical significance tests for information retrieval evaluation. Journal of the Association for Information Science and Technology. https://​doi.​org/​10.​1002/​asi.​24203
Zurück zum Zitat Park, S.-T., Chu, W., Park, S.T., & Chu, W. (2009). Pairwise preference regression for cold-start recommendation. In Proceedings of the 3rd ACM conference on recommender systems, RecSys ’09 (pp. 21–28). ACM, New York, NY, USA. https://doi.org/10.1145/1639714.1639720. Park, S.-T., Chu, W., Park, S.T., & Chu, W. (2009). Pairwise preference regression for cold-start recommendation. In Proceedings of the 3rd ACM conference on recommender systems, RecSys ’09 (pp. 21–28). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​1639714.​1639720.
Zurück zum Zitat Rendle, S., Freudenthaler, C., Gantner, Z., & Schmidt-Thieme, L. (2009). BPR: Bayesian personalized ranking from implicit feedback. In Proceedings of the twenty-fifth conference on uncertainty in artificial intelligence, UAI ’09 (pp. 452–461). AUAI Press, Arlington, VA, US. Rendle, S., Freudenthaler, C., Gantner, Z., & Schmidt-Thieme, L. (2009). BPR: Bayesian personalized ranking from implicit feedback. In Proceedings of the twenty-fifth conference on uncertainty in artificial intelligence, UAI ’09 (pp. 452–461). AUAI Press, Arlington, VA, US.
Zurück zum Zitat Rossetti, M., Stella, F., & Zanker, M. (2016). Contrasting offline and online results when evaluating recommendation algorithms. In Proceedings of the 10th ACM conference on recommender systems, RecSys ’16 (pp. 31–34). ACM, New York, NY, USA. https://doi.org/10.1145/2959100.2959176. Rossetti, M., Stella, F., & Zanker, M. (2016). Contrasting offline and online results when evaluating recommendation algorithms. In Proceedings of the 10th ACM conference on recommender systems, RecSys ’16 (pp. 31–34). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​2959100.​2959176.
Zurück zum Zitat Said, A., & Bellogín, A. (2014). Comparative recommender system evaluation: Benchmarking recommendation frameworks. In Proceedings of the 8th ACM conference on recommender systems, RecSys ’14 (pp. 129–136). ACM, New York, NY, USA. https://doi.org/10.1145/2645710.2645746. Said, A., & Bellogín, A. (2014). Comparative recommender system evaluation: Benchmarking recommendation frameworks. In Proceedings of the 8th ACM conference on recommender systems, RecSys ’14 (pp. 129–136). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​2645710.​2645746.
Zurück zum Zitat Sakai, T. (2006). Evaluating evaluation metrics based on the bootstrap. In Proceedings of the 29th annual international ACM SIGIR conference on research and development in information retrieval, SIGIR ’06 (pp. 525–532). ACM, New York, NY, USA. https://doi.org/10.1145/1148170.1148261. Sakai, T. (2006). Evaluating evaluation metrics based on the bootstrap. In Proceedings of the 29th annual international ACM SIGIR conference on research and development in information retrieval, SIGIR ’06 (pp. 525–532). ACM, New York, NY, USA. https://​doi.​org/​10.​1145/​1148170.​1148261.
Zurück zum Zitat Schnabel, T., Swaminathan, A., Singh, A., Chandak, N., & Joachims, T. (2016). Recommendations as treatments: Debiasing learning and evaluation. In Proceedings of the 33rd international conference on machine learning, ICML ’16 (pp. 1670–1679). Schnabel, T., Swaminathan, A., Singh, A., Chandak, N., & Joachims, T. (2016). Recommendations as treatments: Debiasing learning and evaluation. In Proceedings of the 33rd international conference on machine learning, ICML ’16 (pp. 1670–1679).
Zurück zum Zitat Siroker, D., & Koomen, P. (2013). A/B testing: The most powerful way to turn clicks into customers. New York: Wiley. Siroker, D., & Koomen, P. (2013). A/B testing: The most powerful way to turn clicks into customers. New York: Wiley.
Zurück zum Zitat Spärck Jones, K., & Van Rijsbergen, C.J. (1975). Report on the need for and provision of an ideal information retrieval test collection. British Library Research and Development Reports. University Computer Laboratory Spärck Jones, K., & Van Rijsbergen, C.J. (1975). Report on the need for and provision of an ideal information retrieval test collection. British Library Research and Development Reports. University Computer Laboratory
Zurück zum Zitat Steck, H. (2010). Training and testing of recommender systems on data missing not at random. In Proceedings of the 16th ACM SIGKDD international conference on knowledge discovery and data mining, KDD ’10 (pp. 713–722). New York, NY: ACM. https://doi.org/10.1145/1835804.1835895. Steck, H. (2010). Training and testing of recommender systems on data missing not at random. In Proceedings of the 16th ACM SIGKDD international conference on knowledge discovery and data mining, KDD ’10 (pp. 713–722). New York, NY: ACM. https://​doi.​org/​10.​1145/​1835804.​1835895.
Zurück zum Zitat Swaminathan, A., Krishnamurthy, A., Agarwal, A., Dudík, M., Langford, J., Jose, D., Zitouni, I. (2017). Off-policy evaluation for slate recommendation. In Proceedings of the 31st annual conference on neural information processing systems, NIPS ’17 (pp. 3635–3645). Swaminathan, A., Krishnamurthy, A., Agarwal, A., Dudík, M., Langford, J., Jose, D., Zitouni, I. (2017). Off-policy evaluation for slate recommendation. In Proceedings of the 31st annual conference on neural information processing systems, NIPS ’17 (pp. 3635–3645).
Zurück zum Zitat Valcarce, D., Bellogín, A., Parapar, J., & Castells, P. (2018). On the robustness and discriminative power of information retrieval metrics for top-N recommendation. In Proceedings of the 12th ACM conference on recommender systems, RecSys ’18 (pp. 260–268). New York, NY: ACM. https://doi.org/10.1145/3240323.3240347. Valcarce, D., Bellogín, A., Parapar, J., & Castells, P. (2018). On the robustness and discriminative power of information retrieval metrics for top-N recommendation. In Proceedings of the 12th ACM conference on recommender systems, RecSys ’18 (pp. 260–268). New York, NY: ACM. https://​doi.​org/​10.​1145/​3240323.​3240347.
Zurück zum Zitat Valcarce, D., Parapar, J., & Barreiro, Á. (2016). Efficient pseudo-relevance feedback methods for collaborative filtering recommendation. In Proceedings of the 38th European conference on information retrieval, ECIR ’16 (pp. 602–613). Berlin: Springer. https://doi.org/10.1007/978-3-319-30671-1_44. Valcarce, D., Parapar, J., & Barreiro, Á. (2016). Efficient pseudo-relevance feedback methods for collaborative filtering recommendation. In Proceedings of the 38th European conference on information retrieval, ECIR ’16 (pp. 602–613). Berlin: Springer. https://​doi.​org/​10.​1007/​978-3-319-30671-1_​44.
Zurück zum Zitat Voorhees, E.M. (2002). The philosophy of information retrieval evaluation. In Evaluation of cross-language information retrieval systems: Second workshop of the cross-language evaluation forum, CLEF 2001 (pp. 355–370). Berlin: Springer. https://doi.org/10.1007/3-540-45691-0_34. Voorhees, E.M. (2002). The philosophy of information retrieval evaluation. In Evaluation of cross-language information retrieval systems: Second workshop of the cross-language evaluation forum, CLEF 2001 (pp. 355–370). Berlin: Springer. https://​doi.​org/​10.​1007/​3-540-45691-0_​34.
Zurück zum Zitat Voorhees, E.M. (2005). Overview of the TREC 2004 Robust Track. In ACM SIGIR forum Voorhees, E.M. (2005). Overview of the TREC 2004 Robust Track. In ACM SIGIR forum
Zurück zum Zitat Wang, J., de Vries, A.P., & Reinders, M. J. T. (2006). A user-item relevance model for log-based collaborative filtering. In Proceedings of the 28th European conference on IR research, ECIR ’06 (pp. 37–48). London: Springer. https://doi.org/10.1007/11735106_5. Wang, J., de Vries, A.P., & Reinders, M. J. T. (2006). A user-item relevance model for log-based collaborative filtering. In Proceedings of the 28th European conference on IR research, ECIR ’06 (pp. 37–48). London: Springer. https://​doi.​org/​10.​1007/​11735106_​5.
Zurück zum Zitat Webber, W., Moffat, A., & Zobel, J. (2010). The effect of pooling and evaluation depth on metric stability. In Proceedings of the 3rd international workshop on evaluating information access, EVIA ’10 (pp. 7–15). Webber, W., Moffat, A., & Zobel, J. (2010). The effect of pooling and evaluation depth on metric stability. In Proceedings of the 3rd international workshop on evaluating information access, EVIA ’10 (pp. 7–15).
Zurück zum Zitat Yang, L., Cui, Y., Xuan, Y., Wang, C., Belongie, S., & Estrin, D. (2018). Unbiased offline recommender evaluation for missing-not-at-random implicit feedback. In Proceedings of the 12th ACM conference on recommender systems, RecSys ’18 (pp. 279–287). ACM. Yang, L., Cui, Y., Xuan, Y., Wang, C., Belongie, S., & Estrin, D. (2018). Unbiased offline recommender evaluation for missing-not-at-random implicit feedback. In Proceedings of the 12th ACM conference on recommender systems, RecSys ’18 (pp. 279–287). ACM.
Metadaten
Titel
Assessing ranking metrics in top-N recommendation
verfasst von
Daniel Valcarce
Alejandro Bellogín
Javier Parapar
Pablo Castells
Publikationsdatum
08.06.2020
Verlag
Springer Netherlands
Erschienen in
Discover Computing / Ausgabe 4/2020
Print ISSN: 2948-2984
Elektronische ISSN: 2948-2992
DOI
https://doi.org/10.1007/s10791-020-09377-x

Premium Partner