Skip to main content

2011 | OriginalPaper | Buchkapitel

Decision Tree Modeling for Ranking Data

verfasst von : Philip L. H. Yu, Wai Ming Wan, Paul H. Lee

Erschienen in: Preference Learning

Verlag: Springer Berlin Heidelberg

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Ranking/preference data arises from many applications in marketing, psychology, and politics. We establish a new decision tree model for the analysis of ranking data by adopting the concept of classification and regression tree. The existing splitting criteria are modified in a way that allows them to precisely measure the impurity of a set of ranking data. Two types of impurity measures for ranking data are introduced, namelyg-wise and top-k measures. Theoretical results show that the new measures exhibit properties of impurity functions. In model assessment, the area under the ROC curve (AUC) is applied to evaluate the tree performance. Experiments are carried out to investigate the predictive performance of the tree model for complete and partially ranked data and promising results are obtained. Finally, a real-world application of the proposed methodology to analyze a set of political rankings data is presented.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
2.
Zurück zum Zitat A.P. Bradley, The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 30, 1145–1159 (1997)CrossRef A.P. Bradley, The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 30, 1145–1159 (1997)CrossRef
3.
Zurück zum Zitat L. Breiman, J.H. Friedman, R.A. Olshen, C.J. Stone.Classification and Regression Trees (Belmont, California: Wadsworth, 1984) L. Breiman, J.H. Friedman, R.A. Olshen, C.J. Stone.Classification and Regression Trees (Belmont, California: Wadsworth, 1984)
4.
Zurück zum Zitat W. Cheng, J. Hühn, E. Hüllermeier, Decision tree and instance-based learning for label ranking, inProceedings of the 26th International Conference on Machine Learning (ICML 2009) (Montreal, Canada, 2009) W. Cheng, J. Hühn, E. Hüllermeier, Decision tree and instance-based learning for label ranking, inProceedings of the 26th International Conference on Machine Learning (ICML 2009) (Montreal, Canada, 2009)
5.
Zurück zum Zitat P.A. Chou, Optimal partitioning for classification and regression trees. IEEE Trans. Pattern Anal. Mach. Intell. 13, 340–354 (1991)CrossRef P.A. Chou, Optimal partitioning for classification and regression trees. IEEE Trans. Pattern Anal. Mach. Intell. 13, 340–354 (1991)CrossRef
7.
Zurück zum Zitat C. Drummond, R.C. Holte, What ROC curves can’t do (and cost curves can), inProceedings of the 1st Workshop on ROC Analysis in AI (Valencia, Spain, 2004), pp. 19-Ű26 C. Drummond, R.C. Holte, What ROC curves can’t do (and cost curves can), inProceedings of the 1st Workshop on ROC Analysis in AI (Valencia, Spain, 2004), pp. 19-Ű26
8.
Zurück zum Zitat R.M. Duch, M.A. Taylor, Postmaterialism and the economic condition. Am. J. Pol. Sci. 37, 747–778 (1993)CrossRef R.M. Duch, M.A. Taylor, Postmaterialism and the economic condition. Am. J. Pol. Sci. 37, 747–778 (1993)CrossRef
9.
Zurück zum Zitat J. Fürnkranz, E. Hüllermeier, Pairwise preference learning and ranking, inProceedings of the 14th European Conference on Machine Learning (ECML-03) (Springer, Cavtat, Croatia, 2003), pp. 145–156 J. Fürnkranz, E. Hüllermeier, Pairwise preference learning and ranking, inProceedings of the 14th European Conference on Machine Learning (ECML-03) (Springer, Cavtat, Croatia, 2003), pp. 145–156
10.
Zurück zum Zitat P. Geurts, L. Wehenkel, A. Florence, Kernelizing the output of tree-based methods, inProceedings of the 23rd International Conference on Machine Learning (ICML-06), (Pittsburgh, Pennsylvania, 2006), pp. 345–352 P. Geurts, L. Wehenkel, A. Florence, Kernelizing the output of tree-based methods, inProceedings of the 23rd International Conference on Machine Learning (ICML-06), (Pittsburgh, Pennsylvania, 2006), pp. 345–352
11.
Zurück zum Zitat D.J. Hand, R.J. Till, A simple generalisation of the area under the ROC curve for multiple class classification problems. Mach. Learn. 45(2), 171–186 (2001)MATHCrossRef D.J. Hand, R.J. Till, A simple generalisation of the area under the ROC curve for multiple class classification problems. Mach. Learn. 45(2), 171–186 (2001)MATHCrossRef
12.
Zurück zum Zitat J.A. Hanley, B.J. McNeil, The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 143, 29–36 (1982) J.A. Hanley, B.J. McNeil, The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 143, 29–36 (1982)
13.
Zurück zum Zitat E. Hüllermeier, J. Fürnkranz, On Minimizing the Position Error in Label Ranking, inProceedings of the 17th European Conference on Machine Learning (ECML-07) (Springer, Warsawa, Poland, 2007), pp. 583–590 E. Hüllermeier, J. Fürnkranz, On Minimizing the Position Error in Label Ranking, inProceedings of the 17th European Conference on Machine Learning (ECML-07) (Springer, Warsawa, Poland, 2007), pp. 583–590
14.
Zurück zum Zitat E. Hüllermeier, J. Fürnkranz, W. Cheng, K. Brinker, Label ranking by learning pairwise preferences. Artif. Intell. 172(16–17), 1897–1916 (2008)MATHCrossRef E. Hüllermeier, J. Fürnkranz, W. Cheng, K. Brinker, Label ranking by learning pairwise preferences. Artif. Intell. 172(16–17), 1897–1916 (2008)MATHCrossRef
15.
Zurück zum Zitat R. Inglehart,The Silent Revolution: Changing Values and Political Styles among Western Publics (Princeton Univerity Press, Princeton, 1977) R. Inglehart,The Silent Revolution: Changing Values and Political Styles among Western Publics (Princeton Univerity Press, Princeton, 1977)
16.
Zurück zum Zitat R. Jowell, L. Brook, L. Dowds,International Social Attributes: the 10th BSA Report (Dartmouth Publishing, Aldershot, 1993) R. Jowell, L. Brook, L. Dowds,International Social Attributes: the 10th BSA Report (Dartmouth Publishing, Aldershot, 1993)
17.
Zurück zum Zitat M.G. Karlaftis, Predicting mode choice through multivariate recursive partitioning. J. Trans. Eng. 130(22), 245–250 (2004)CrossRef M.G. Karlaftis, Predicting mode choice through multivariate recursive partitioning. J. Trans. Eng. 130(22), 245–250 (2004)CrossRef
18.
Zurück zum Zitat G.V. Kass, An exploratory technique for investigation large quantities of categorical data. Appl. Stat. 29, 119–127 (1980)CrossRef G.V. Kass, An exploratory technique for investigation large quantities of categorical data. Appl. Stat. 29, 119–127 (1980)CrossRef
19.
Zurück zum Zitat W.-Y. Loh, Y.-S. Shih, Split selection methods for classification trees. Statistica Sinica 7, 815–840 (1997)MathSciNetMATH W.-Y. Loh, Y.-S. Shih, Split selection methods for classification trees. Statistica Sinica 7, 815–840 (1997)MathSciNetMATH
20.
Zurück zum Zitat J.I. Marden, Analyzing and Modeling Rank Data (Chapman & Hall, 1995) J.I. Marden, Analyzing and Modeling Rank Data (Chapman & Hall, 1995)
21.
Zurück zum Zitat J.R. Quinlan, Induction of decision trees. Mach. Learn. 1, 81–106 (1986) J.R. Quinlan, Induction of decision trees. Mach. Learn. 1, 81–106 (1986)
22.
Zurück zum Zitat J.R. Quinlan.C4.5: Programs for Machine Learning (Morgan Kaufmann, 1993) J.R. Quinlan.C4.5: Programs for Machine Learning (Morgan Kaufmann, 1993)
23.
Zurück zum Zitat B.D. Ripley,Pattern Recognition and Neural Networks (Cambridge University Press, Cambridge, 1996)MATH B.D. Ripley,Pattern Recognition and Neural Networks (Cambridge University Press, Cambridge, 1996)MATH
24.
Zurück zum Zitat L. Rokach, O. Maimon, Decision trees, inThe Data Mining and Knowledge Discovery Handbook, ed. by O. Maimon, L. Rokach (Springer, Berlin, 2005), pp. 165–192CrossRef L. Rokach, O. Maimon, Decision trees, inThe Data Mining and Knowledge Discovery Handbook, ed. by O. Maimon, L. Rokach (Springer, Berlin, 2005), pp. 165–192CrossRef
25.
Zurück zum Zitat R. Siciliano, F. Mola, Multivariate data analysis and modeling through classification and regression trees. Comput. Stat. Data Anal. 32, 285–301 (2000)MathSciNetMATHCrossRef R. Siciliano, F. Mola, Multivariate data analysis and modeling through classification and regression trees. Comput. Stat. Data Anal. 32, 285–301 (2000)MathSciNetMATHCrossRef
Metadaten
Titel
Decision Tree Modeling for Ranking Data
verfasst von
Philip L. H. Yu
Wai Ming Wan
Paul H. Lee
Copyright-Jahr
2011
Verlag
Springer Berlin Heidelberg
DOI
https://doi.org/10.1007/978-3-642-14125-6_5

Premium Partner