Skip to main content
Top

2017 | OriginalPaper | Chapter

Efficient Top Rank Optimization with Gradient Boosting for Supervised Anomaly Detection

Authors : Jordan Frery, Amaury Habrard, Marc Sebban, Olivier Caelen, Liyun He-Guelton

Published in: Machine Learning and Knowledge Discovery in Databases

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

In this paper we address the anomaly detection problem in a supervised setting where positive examples might be very sparse. We tackle this task with a learning to rank strategy by optimizing a differentiable smoothed surrogate of the so-called Average Precision (AP). Despite its non-convexity, we show how to use it efficiently in a stochastic gradient boosting framework. We show that using AP is much better to optimize the top rank alerts than the state of the art measures. We demonstrate on anomaly detection tasks that the interest of our method is even reinforced in highly unbalanced scenarios.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Footnotes
2
ATOS/Wolrdline is leader in e-transaction payments http://​worldline.​com/​.
 
3
Note that we did not use Adarank in our evaluation because the weights updates rely on a notion of query that is not adapted to our framework.
 
Literature
1.
go back to reference Cortes, C., Mohri, M.: AUC optimization vs. error rate minimization. In: NIPS, pp. 313–320 (2003) Cortes, C., Mohri, M.: AUC optimization vs. error rate minimization. In: NIPS, pp. 313–320 (2003)
2.
go back to reference Chawla, N.V., Bowyer, K.W., Hall, L.O., Kegelmeyer, W.P.: SMOTE: synthetic minority over-sampling technique. JAIR 16, 321–357 (2002)MATH Chawla, N.V., Bowyer, K.W., Hall, L.O., Kegelmeyer, W.P.: SMOTE: synthetic minority over-sampling technique. JAIR 16, 321–357 (2002)MATH
3.
go back to reference Ramentol, E., Caballero, Y., Bello, R., Herrera, F.: SMOTE-RSB*: a hybrid preprocessing approach based on oversampling and undersampling for high imbalanced data-sets using SMOTE and rough sets theory. Knowl. Inf. Syst. 33, 245–265 (2012)CrossRef Ramentol, E., Caballero, Y., Bello, R., Herrera, F.: SMOTE-RSB*: a hybrid preprocessing approach based on oversampling and undersampling for high imbalanced data-sets using SMOTE and rough sets theory. Knowl. Inf. Syst. 33, 245–265 (2012)CrossRef
5.
go back to reference Seiffert, C., Khoshgoftaar, T.M., Hulse, J.V., Napolitano, A.: RUSBoost: a hybrid approach to alleviating class imbalance. IEEE Trans. Syst. Man Cybern. 40(1), 185–197 (2010)CrossRef Seiffert, C., Khoshgoftaar, T.M., Hulse, J.V., Napolitano, A.: RUSBoost: a hybrid approach to alleviating class imbalance. IEEE Trans. Syst. Man Cybern. 40(1), 185–197 (2010)CrossRef
6.
go back to reference Fan, W., Stolfo, S.J., Zhang, J., Chan, P.K.: AdaCost: misclassification cost-sensitive boosting. In: ICML, pp. 97–105 (1999) Fan, W., Stolfo, S.J., Zhang, J., Chan, P.K.: AdaCost: misclassification cost-sensitive boosting. In: ICML, pp. 97–105 (1999)
7.
8.
go back to reference Niculescu-Mizil, A., Caruana, R.: Predicting good probabilities with supervised learning. In: ICML, pp. 625–632 (2005) Niculescu-Mizil, A., Caruana, R.: Predicting good probabilities with supervised learning. In: ICML, pp. 625–632 (2005)
9.
go back to reference Dal Pozzolo, A., Caelen, O., Johnson, R.A., Bontempi, G.: Calibrating probability with undersampling for unbalanced classification. In: SSCI, pp. 159–166 (2015) Dal Pozzolo, A., Caelen, O., Johnson, R.A., Bontempi, G.: Calibrating probability with undersampling for unbalanced classification. In: SSCI, pp. 159–166 (2015)
11.
go back to reference Joachims, T.: Optimizing search engines using clickthrough data. In: SIGKDD, pp. 133–142 (2002) Joachims, T.: Optimizing search engines using clickthrough data. In: SIGKDD, pp. 133–142 (2002)
12.
go back to reference Yue, Y., Finley, T., Radlinski, F., Joachims, T.: A support vector method for optimizing average precision. In: SIGIR, pp. 271–278 (2007) Yue, Y., Finley, T., Radlinski, F., Joachims, T.: A support vector method for optimizing average precision. In: SIGIR, pp. 271–278 (2007)
14.
go back to reference Freund, Y., Schapire, R., Abe, N.: A short introduction to boosting. J.-Jpn Soc. Artif. Intell. 14(771–780), 1612 (1999) Freund, Y., Schapire, R., Abe, N.: A short introduction to boosting. J.-Jpn Soc. Artif. Intell. 14(771–780), 1612 (1999)
15.
go back to reference Chapelle, O., Chang, Y.: Yahoo! learning to rank challenge overview, pp. 1–24 (2011) Chapelle, O., Chang, Y.: Yahoo! learning to rank challenge overview, pp. 1–24 (2011)
17.
go back to reference Freund, Y., Iyer, R., Schapire, R.E., Singer, Y.: An efficient boosting algorithm for combining preferences. J. Mach. Learn. Res. 4, 933–969 (2003)MathSciNetMATH Freund, Y., Iyer, R., Schapire, R.E., Singer, Y.: An efficient boosting algorithm for combining preferences. J. Mach. Learn. Res. 4, 933–969 (2003)MathSciNetMATH
18.
go back to reference Burges, C., Shaked, T., Renshaw, E., Lazier, A., Deeds, M., Hamilton, N., Hullender, G.: Learning to rank using gradient descent. In: ICML, pp. 89–96 (2005) Burges, C., Shaked, T., Renshaw, E., Lazier, A., Deeds, M., Hamilton, N., Hullender, G.: Learning to rank using gradient descent. In: ICML, pp. 89–96 (2005)
19.
go back to reference Herschtal, A., Raskutti, B.: Optimising area under the ROC curve using gradient descent. In: Proceedings of the Twenty-First International Conference on Machine learning, p. 49. ACM (2004) Herschtal, A., Raskutti, B.: Optimising area under the ROC curve using gradient descent. In: Proceedings of the Twenty-First International Conference on Machine learning, p. 49. ACM (2004)
20.
go back to reference Burges, C.J.: From RankNet to LambdaRank to LambdaMART: an overview. Learning 11, 23–581 (2010) Burges, C.J.: From RankNet to LambdaRank to LambdaMART: an overview. Learning 11, 23–581 (2010)
21.
go back to reference Xu, J., Li, H.: AdaRank: a boosting algorithm for information retrieval. In: SIGIR, pp. 391–398. ACM (2007) Xu, J., Li, H.: AdaRank: a boosting algorithm for information retrieval. In: SIGIR, pp. 391–398. ACM (2007)
22.
go back to reference Wu, Q., Burges, C.J., Svore, K.M., Gao, J.: Adapting boosting for information retrieval measures. Inf. Retr. 13(3), 254–270 (2010)CrossRef Wu, Q., Burges, C.J., Svore, K.M., Gao, J.: Adapting boosting for information retrieval measures. Inf. Retr. 13(3), 254–270 (2010)CrossRef
23.
go back to reference Burges, C.J., Ragno, R., Le, Q.V.: Learning to rank with nonsmooth cost functions. In: Schölkopf, P.B., Platt, J.C., Hoffman, T., (eds.) NIPS, pp. 193–200 (2007) Burges, C.J., Ragno, R., Le, Q.V.: Learning to rank with nonsmooth cost functions. In: Schölkopf, P.B., Platt, J.C., Hoffman, T., (eds.) NIPS, pp. 193–200 (2007)
24.
go back to reference Hanley, J.A., McNeil, B.J.: The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 143(1), 29–36 (1982)CrossRef Hanley, J.A., McNeil, B.J.: The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 143(1), 29–36 (1982)CrossRef
26.
go back to reference Li, N., Jin, R., Zhou, Z.H.: Top rank optimization in linear time. In: NIPS, pp. 1502–1510 (2014) Li, N., Jin, R., Zhou, Z.H.: Top rank optimization in linear time. In: NIPS, pp. 1502–1510 (2014)
Metadata
Title
Efficient Top Rank Optimization with Gradient Boosting for Supervised Anomaly Detection
Authors
Jordan Frery
Amaury Habrard
Marc Sebban
Olivier Caelen
Liyun He-Guelton
Copyright Year
2017
DOI
https://doi.org/10.1007/978-3-319-71249-9_2

Premium Partner