skip to main content
10.1145/1718487.1718509acmconferencesArticle/Chapter ViewAbstractPublication PageswsdmConference Proceedingsconference-collections
research-article

Improving quality of training data for learning to rank using click-through data

Published:04 February 2010Publication History

ABSTRACT

In information retrieval, relevance of documents with respect to queries is usually judged by humans, and used in evaluation and/or learning of ranking functions. Previous work has shown that certain level of noise in relevance judgments has little effect on evaluation, especially for comparison purposes. Recently learning to rank has become one of the major means to create ranking models in which the models are automatically learned from the data derived from a large number of relevance judgments. As far as we know, there was no previous work about quality of training data for learning to rank, and this paper tries to study the issue. Specifically, we address three problems. Firstly, we show that the quality of training data labeled by humans has critical impact on the performance of learning to rank algorithms. Secondly, we propose detecting relevance judgment errors using click-through data accumulated at a search engine. Two discriminative models, referred to as sequential dependency model and full dependency model, are proposed to make the detection. Both models consider the conditional dependency of relevance labels and thus are more powerful than the conditionally independent model previously proposed for other tasks. Finally, we verify that using training data in which the errors are detected and corrected by our method, we can improve the performance of learning to rank algorithms.

References

  1. R. Agrawal, A. Halverson, K. Kenthapadi, N. Mishra, and P. Tsaparas. Generating labels from clicks. In WSDM '09: Proceedings of the Second ACM International Conference on Web Search and Data Mining, pages 172--181. ACM, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. P. Bailey, N. Craswell, I. Soboroff, P. Thomas, A.P. de Vries, and E. Yilmaz. Relevance assessment: are judges exchangeable and does it matter. In SIGIR '08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval, pages 667--674. ACM, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. B. Carterette and R. Jones. Evaluating search engines by modeling the relationship between relevance and clicks. In Advances in Neural Information Processing Systems, volume 20, 2007.Google ScholarGoogle Scholar
  4. N. Craswell and M. Szummer. Random walks on the click graph. In SIGIR '07: Proceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval, pages 239--246. ACM, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Y. Freund, R. Iyer, R.E. Schapire, and Y. Singer. An efficient boosting algorithm for combining preferences. J. Mach. Learn. Res., 4:933--969, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. S.P. Harter. Variations in relevance assessments and the measurement of retrieval effectiveness. J. Am. Soc. Inf. Sci., 47(1):37--49, 1996. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. S. Ji, K. Zhou, C. Liao, Z. Zheng, G.-R. Xue, O. Chapelle, G. Sun, and H. Zha. Global ranking by exploiting user clicks. In SIGIR '09: Proceedings of the 32nd international ACM SIGIR conference on Research and development in information retrieval, pages 35--42. ACM, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. T. Joachims. Optimizing search engines using clickthrough data. In KDD '02: Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining, pages 133--142. ACM, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. T. Joachims, L. Granka, B. Pan, H. Hembrooke, and G. Gay. Accurately interpreting clickthrough data as implicit feedback. In SIGIR '05: Proceedings of the 28th annual international ACM SIGIR conference on Research and development in information retrieval, pages 154--161. ACM, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. J. Kamps, M. Koolen, and A. Trotman. Comparative analysis of clicks and judgments for ir evaluation. In WSCD '09: Proceedings of the 2009 workshop on Web Search Click Data, pages 80--87. ACM, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. J.D. Lafferty, A. McCallum, and F.C.N. Pereira. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In ICML '01: Proceedings of the Eighteenth International Conference on Machine Learning, pages 282--289. Morgan Kaufmann Publishers Inc., 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. C. Macdonald and I. Ounis. Usefulness of quality click-through data for training. In WSCD '09: Proceedings of the 2009 workshop on Web Search Click Data, pages 75--79. ACM, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Q. Mei, D. Zhou, and K. Church. Query suggestion using hitting time. In CIKM '08: Proceeding of the 17th ACM conference on Information and knowledge management, pages 469--478. ACM, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. D. Metzler and W.B. Croft. A markov random field model for term dependencies. In SIGIR '05: Proceedings of the 28th annual international ACM SIGIR conference on Research and development in information retrieval, pages 472--479. ACM, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. F. Qiu and J. Cho. Automatic identification of user interest for personalized search. In WWW '06: Proceedings of the 15th international conference on World Wide Web, pages 727--736. ACM, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. F. Radlinski and T. Joachims. Query chains: learning to rank from implicit feedback. In KDD '05: Proceedings of the eleventh ACM SIGKDD international conference on Knowledge discovery in data mining, pages 239--248. ACM, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. F. Radlinski, M. Kurup, and T. Joachims. How does clickthrough data reflect retrieval quality? In CIKM'08: Proceeding of the 17th ACM conference on Information and knowledge management, pages 43--52. ACM, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. P. Ravikumar and J. Lafferty. Quadratic programming relaxations for metric labeling and markov random field map estimation. In ICML '06: Proceedings of the 23rd international conference on Machine learning, pages 737--744. ACM, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. E. Voorhees and D. Harman. Overview of the fifth text retrieval conference (trec-5). In Proceedings of the sixth Text REtrieval Conference (TREC-6). (NIST Special Publication 500--240, 1997.Google ScholarGoogle Scholar
  20. E.M. Voorhees. Variations in relevance judgments and the measurement of retrieval effectiveness. In SIGIR'98: Proceedings of the 21st annual international ACM SIGIR conference on Research and development in information retrieval, pages 315--323. ACM, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. J. Xu and H. Li. Adarank: a boosting algorithm for information retrieval. In SIGIR '07: Proceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval, pages 391--398. ACM, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Y. Yue, T. Finley, F. Radlinski, and T. Joachims. A support vector method for optimizing average precision. In SIGIR '07: Proceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval, pages 271--278. ACM, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Improving quality of training data for learning to rank using click-through data

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      WSDM '10: Proceedings of the third ACM international conference on Web search and data mining
      February 2010
      468 pages
      ISBN:9781605588896
      DOI:10.1145/1718487

      Copyright © 2010 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 4 February 2010

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      Overall Acceptance Rate498of2,863submissions,17%

      Upcoming Conference

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader